10,000 Matching Annotations
  1. Sep 2024
    1. Reviewer #1 (Public Review):

      Summary:

      Zhou and colleagues developed a computational model of replay that heavily builds on cognitive models of memory in context (e.g., the context-maintenance and retrieval model), which have been successfully used to explain memory phenomena in the past. Their model produces results that mirror previous empirical findings in rodents and offers a new computational framework for thinking about replay.

      Strengths:

      The model is compelling and seems to explain a number of findings from the rodent literature. It is commendable that the authors implement commonly used algorithms from wakefulness to model sleep/rest, thereby linking wake and sleep phenomena in a parsimonious way. Additionally, the manuscript's comprehensive perspective on replay, bridging humans and non-human animals, enhanced its theoretical contribution.

      Weaknesses:

      This reviewer is not a computational neuroscientist by training, so some comments may stem from misunderstandings. I hope the authors would see those instances as opportunities to clarify their findings for broader audiences.

      (1) The model predicts that temporally close items will be co-reactivated, yet evidence from humans suggests that temporal context doesn't guide sleep benefits (instead, semantic connections seem to be of more importance; Liu and Ranganath 2021, Schechtman et al 2023). Could these findings be reconciled with the model or is this a limitation of the current framework?

      (2) During replay, the model is set so that the next reactivated item is sampled without replacement (i.e., the model cannot get "stuck" on a single item). I'm not sure what the biological backing behind this is and why the brain can't reactivate the same item consistently. Furthermore, I'm afraid that such a rule may artificially generate sequential reactivation of items regardless of wake training. Could the authors explain this better or show that this isn't the case?

      (3) If I understand correctly, there are two ways in which novelty (i.e., less exposure) is accounted for in the model. The first and more talked about is the suppression mechanism (lines 639-646). The second is a change in learning rates (lines 593-595). It's unclear to me why both procedures are needed, how they differ, and whether these are two different mechanisms that the model implements. Also, since the authors controlled the extent to which each item was experienced during wakefulness, it's not entirely clear to me which of the simulations manipulated novelty on an individual item level, as described in lines 593-595 (if any).

      As to the first mechanism - experience-based suppression - I find it challenging to think of a biological mechanism that would achieve this and is selectively activated immediately before sleep (somehow anticipating its onset). In fact, the prominent synaptic homeostasis hypothesis suggests that such suppression, at least on a synaptic level, is exactly what sleep itself does (i.e., prune or weaken synapses that were enhanced due to learning during the day). This begs the question of whether certain sleep stages (or ultradian cycles) may be involved in pruning, whereas others leverage its results for reactivation (e.g., a sequential hypothesis; Rasch & Born, 2013). That could be a compelling synthesis of this literature. Regardless of whether the authors agree, I believe that this point is a major caveat to the current model. It is addressed in the discussion, but perhaps it would be beneficial to explicitly state to what extent the results rely on the assumption of a pre-sleep suppression mechanism.

      (4) As the manuscript mentions, the only difference between sleep and wake in the model is the initial conditions (a0). This is an obvious simplification, especially given the last author's recent models discussing the very different roles of REM vs NREM. Could the authors suggest how different sleep stages may relate to the model or how it could be developed to interact with other successful models such as the ones the last author has developed (e.g., C-HORSE)? Finally, I wonder how the model would explain findings (including the authors') showing a preference for reactivation of weaker memories. The literature seems to suggest that it isn't just a matter of novelty or exposure, but encoding strength. Can the model explain this? Or would it require additional assumptions or some mechanism for selective endogenous reactivation during sleep and rest?

      (5) Lines 186-200 - Perhaps I'm misunderstanding, but wouldn't it be trivial that an external cue at the end-item of Figure 7a would result in backward replay, simply because there is no potential for forward replay for sequences starting at the last item (there simply aren't any subsequent items)? The opposite is true, of course, for the first-item replay, which can't go backward. More generally, my understanding of the literature on forward vs backward replay is that neither is linked to the rodent's location. Both commonly happen at a resting station that is further away from the track. It seems as though the model's result may not hold if replay occurs away from the track (i.e. if a0 would be equal for both pre- and post-run).

      (6) The manuscript describes a study by Bendor & Wilson (2012) and tightly mimics their results. However, notably, that study did not find triggered replay immediately following sound presentation, but rather a general bias toward reactivation of the cued sequence over longer stretches of time. In other words, it seems that the model's results don't fully mirror the empirical results. One idea that came to mind is that perhaps it is the R/L context - not the first R/L item - that is cued in this study. This is in line with other TMR studies showing what may be seen as contextual reactivation. If the authors think that such a simulation may better mirror the empirical results, I encourage them to try. If not, however, this limitation should be discussed.

      (7) There is some discussion about replay's benefit to memory. One point of interest could be whether this benefit changes between wake and sleep. Relatedly, it would be interesting to see whether the proportion of forward replay, backward replay, or both correlated with memory benefits. I encourage the authors to extend the section on the function of replay and explore these questions.

      (8) Replay has been mostly studied in rodents, with few exceptions, whereas CMR and similar models have mostly been used in humans. Although replay is considered a good model of episodic memory, it is still limited due to limited findings of sequential replay in humans and its reliance on very structured and inherently autocorrelated items (i.e., place fields). I'm wondering if the authors could speak to the implications of those limitations on the generalizability of their model. Relatedly, I wonder if the model could or does lead to generalization to some extent in a way that would align with the complementary learning systems framework.

    1. But now we have more shows, so we’re still recapping a lot — it’s just that there’s so many more.

      Here the author explains another reason for the decline in TV recaps. I believe that overstimulation plays a huge role in this as well. There are a lot more shows to recap, however we have to manage our time well and refrain from overconsumption.

    2. “The culture of this extreme dissection of TV that recaps started has grown. There are just so many different formats where you can be doing that,” Emami says. At Vulture, recaps are “still a very big part of what we do, but I also think it’s now just one part of what we do. It’s one part of a coverage plan, and that can include explainers, think pieces, what are the biggest questions asked after this episode of Westworld.” Recaps were just one expression of an idea that still holds sway over the internet, and how audiences talk about TV in general: essentially, that it’s worth talking about — publicly, rigorously, and joyfully. As long as that philosophy remains intact, its execution is both flexible and secondary. Netflix shows may not make for good recaps, but they can still spawn a meme like Barb, a perfect fusion of internet weirdos and the unwitting object of their passion that followed the spirit of recaps, if not their letter. The permission to honor something you love by unpacking it, and the idea that affection itself is reason itself for unpacking, is a difficult dam to unburst.

      The passage reflects on the evolution of television criticism and audience engagement, highlighting the importance of discussion and analysis in a variety of formats while emphasizing the joy and affection that drive these conversations.

    3. I don’t think it’s gonna be the dominant form anymore, just because everyone’s watching on their own schedule now,”

      I find that some shows that still release episodically are popular so I believe its still possible.

    4. “I feel like we’re coming back around to an era where recaps could take off again,” he says. “Because I feel like people find the conversation on social media to be increasingly toxic, and want to just read somebody who’s well informed.

      I would say the "The Recap" is still very much alive and thriving. The part of it that has died is it being text based, it's all on video now. Youtube and TikTok are flooded with users doing recaps for tv shows.

    1. But they always do. We look at their guided free-writes together and highlight patterns of connected words or moments of sharp, conceptual thinking. Sometimes it’s an “ah-ha” moment; a small revelation.

      Sometimes us as writers we tend to overthink what we are going to write about, and come to the conclusion that we dont have nothing to write. In reality, we just need to put some words on the paper, and the ideas will start flowing.

    2. We experience a tragedy on two levels: the first is the one that feels, that grieves, that aches in every room where he isn’t; that plans the ceremony for the children, that speaks to the vet on the phone about the bill for the dead dog. The second level is the writer who is gathering evidence, who understands this story will grow more stark with time—who knows it’s too painful to write about yet but will nonetheless hold onto it in fragments, for down the road it will become part of the experience of sensemaking.

      Something like grief is very important in our journey to grow and living because no war was ever won without losses. However, when it comes to unexpected losses, we normally don't comprehend what just happened so sometimes people use coping skills like writing to make sense of what they lost.

    1. Author response:

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      In their paper, Zhan et al. have used Pf genetic data from simulated data and Ghanaian field samples to elucidate a relationship between multiplicity of infection (MOI) (the number of distinct parasite clones in a single host infection) and force of infection (FOI). Specifically, they use sequencing data from the var genes of Pf along with Bayesian modeling to estimate MOI individual infections and use these values along with methods from queueing theory that rely on various assumptions to estimate FOI. They compare these estimates to known FOIs in a simulated scenario and describe the relationship between these estimated FOI values and another commonly used metric of transmission EIR (entomological inoculation rate).

      This approach does fill an important gap in malaria epidemiology, namely estimating the force of infection, which is currently complicated by several factors including superinfection, unknown duration of infection, and highly genetically diverse parasite populations. The authors use a new approach borrowing from other fields of statistics and modeling and make extensive efforts to evaluate their approach under a range of realistic sampling scenarios. However, the write-up would greatly benefit from added clarity both in the description of methods and in the presentation of the results. Without these clarifications, rigorously evaluating whether the author's proposed method of estimating FOI is sound remains difficult. Additionally, there are several limitations that call into question the stated generalizability of this method that should at minimum be further discussed by authors and in some cases require a more thorough evaluation.

      Major comments:

      (1) Description and evaluation of FOI estimation procedure.

      a. The methods section describing the two-moment approximation and accompanying appendix is lacking several important details. Equations on lines 891 and 892 are only a small part of the equations in Choi et al. and do not adequately describe the procedure notably several quantities in those equations are never defined some of them are important to understand the method (e.g. A, S as the main random variables for inter-arrival times and service times, aR and bR which are the known time average quantities, and these also rely on the squared coefficient of variation of the random variable which is also never introduced in the paper). Without going back to the Choi paper to understand these quantities, and to understand the assumptions of this method it was not possible to follow how this works in the paper. At a minimum, all variables used in the equations should be clearly defined. 

      We thank the reviewer for this useful comment. We plan to clarify the method, including all the relevant variables in our revised manuscript. The reviewer is correct in pointing out that there are more sections and equations in Choi et al., including the derivation of an exact expression for the steady-state queue-length distribution and the two-moment approximation for the queue-length distribution. Since only the latter was directly utilized in our work, we included in the first version of our manuscript only material on this section and not the other. We agree with the reviewer on readers benefiting from additional information on the derivation of the exact expression for the steady-state queue-length distribution. Therefore, we will summarize the derivation of this expression in our revised manuscript. Regarding the assumptions of the method we applied, especially those for going from the exact expression to the two-moment approximation, we did describe these in the Materials and Methods of our manuscript. We recognize from this comment that the writing and organization of this information may not have been sufficiently clear. We had separated the information on this method into two parts, with the descriptive summary placed in the Materials and Methods and the equations or mathematical formula placed in the Appendix. This can make it difficult for readers to connect the two parts and remember what was introduced earlier in the Materials and Methods when reading the equations and mathematical details in the Appendix. For our revised manuscript, we plan to cover both parts in the Materials and Methods, and to provide more of the technical details in one place, which will be easier to understand and follow.

      b. Additionally, the description in the main text of how the queueing procedure can be used to describe malaria infections would benefit from a diagram currently as written it's very difficult to follow. 

      We thank the reviewer for this suggestion. We will add a diagram illustrating the connection between the queueing procedure and malaria transmission.

      c. Just observing the box plots of mean and 95% CI on a plot with the FOI estimate (Figures 1, 2, and 10-14) is not sufficient to adequately assess the performance of this estimator. First, it is not clear whether the authors are displaying the bootstrapped 95%CIs or whether they are just showing the distribution of the mean FOI taken over multiple simulations, and then it seems that they are also estimating mean FOI per host on an annual basis. Showing a distribution of those per-host estimates would also be helpful. Second, a more quantitative assessment of the ability of the estimator to recover the truth across simulations (e.g. proportion of simulations where the truth is captured in the 95% CI or something like this) is important in many cases it seems that the estimator is always underestimating the true FOI and may not even contain the true value in the FOI distribution (e.g. Figure 10, Figure 1 under the mid-IRS panel). But it's not possible to conclude one way or the other based on this visualization. This is a major issue since it calls into question whether there is in fact data to support that these methods give good and consistent FOI estimates. 

      There appears to be some confusion on what we display in some key figures. We will clarify this further both here and in the revised text. In Figures 1, 2, and 10-14, we displayed the bootstrapped distributions including the 95% CIs. These figures do not show the distribution of the mean FOI taken over multiple simulations. We estimated mean FOI on an annual basis per host in the following sense. Both of our proposed methods require either a steady-state queue length distribution, or moments of this distribution for FOI inference. However, we only have one realization or observation for each individual host, and we do not have access to either the time-series observation of a single individual’s MOI or many realizations of a single individual’s MOI at the same sampling time. This is typically the case for empirical data, although numerical simulations could circumvent this limitation and generate such output. Nonetheless, we do have a queue length distribution at the population level for both the simulation output and the empirical data, which can be obtained by simply aggregating MOI estimates across all sampled individuals. We use this population-level queue length distribution to represent and approximate the steady-state queue length distribution at the individual level. Such representation or approximation does not consider explicitly any individual heterogeneity due to biology or transmission. The estimated FOI is per host in the sense of representing the FOI experienced by an individual host whose queue length distribution is approximated from the collection of all sampled individuals. The true FOI per host per year in the simulation output is obtained from dividing the total FOI of all hosts per year by the total number of all hosts. Therefore, our estimator, combined with the demographic information on population size, is for the total number of Plasmodium falciparum infections acquired by all individual hosts in the population of interest per year.

      We evaluated the impact of individual heterogeneity on FOI inference by introducing individual heterogeneity into the simulations. With a considerable amount of transmission heterogeneity across individuals (namely 2/3 of the population receiving more than 90% of all bites whereas the remaining 1/3 receives the rest of the bites), our two methods exhibit a similar performance than those of the homogeneous transmission scenarios.

      Concerning the second point, we will add a quantitative assessment of the ability of the estimator to recover the truth across simulations and include this information in the legend of each figure. In particular, we will provide the proportion of simulations where the truth is captured by the entire bootstrap distribution, in addition to some measure of relative deviation, such as the relative difference between the true FOI value and the median of the bootstrap distribution for the estimate. This assessment will be a valuable addition, but please note that the comparisons we have provided in a graphical way do illustrate the ability of the methods to estimate “sensible” values, close to the truth despite multiple sources of errors. “Close” is here relative to the scale of variation of FOI in the field and to the kind of precision that would be useful in an empirical context. From a practical perspective based on the potential range of variation of FOI, the graphical results already illustrate that the estimated distributions would be informative.

      d. Furthermore the authors state in the methods that the choice of mean and variance (and thus second moment) parameters for inter-arrival times are varied widely, however, it's not clear what those ranges are there needs to be a clear table or figure caption showing what combinations of values were tested and which results are produced from them, this is an essential component of the method and it's impossible to fully evaluate its performance without this information. This relates to the issue of selecting the mean and variance values that maximize the likelihood of observing a given distribution of MOI estimates, this is very unclear since no likelihoods have been written down in the methods section of the main text, which likelihood are the authors referring to, is this the probability distribution of the steady state queue length distribution? At other places the authors refer to these quantities as Maximum Likelihood estimators, how do they know they have found the MLE? There are no derivations in the manuscript to support this. The authors should specify the likelihood and include in an appendix an explanation of why their estimation procedure is in fact maximizing this likelihood, preferably with evidence of the shape of the likelihood, and how fine the grid of values they tested is for their mean and variance since this could influence the overall quality of the estimation procedure. 

      We thank the reviewer for pointing out these aspects of the work that can be further clarified. We will specify the ranges for the choice of mean and variance parameters for inter-arrival times as well as the grid of values tested in the corresponding figure caption or in a separate supplementary table. We maximized the likelihood of observing the set of individual MOI estimates in a sampled population given steady queue length distributions (with these distributions based on the two-moment approximation method for different combinations of the mean and variance of inter-arrival times). We will add a section to either the Materials and Methods or the Appendix in our revised manuscript including an explicit formulation of the likelihood.

      We will add example figures on the shape of the likelihood to the Appendix. We will also test how choices of the grid of values influence the overall quality of the estimation procedure. Specifically, we will further refine the grid of values to include more points and examine whether the results of FOI inference are consistent and robust against each other.

      (2) Limitation of FOI estimation procedure.

      a. The authors discuss the importance of the duration of infection to this problem. While I agree that empirically estimating this is not possible, there are other options besides assuming that all 1-5-year-olds have the same duration of infection distribution as naïve adults co-infected with syphilis. E.g. it would be useful to test a wide range of assumed infection duration and assess their impact on the estimation procedure. Furthermore, if the authors are going to stick to the described method for duration of infection, the potentially limited generalizability of this method needs to be further highlighted in both the introduction, and the discussion. In particular, for an estimated mean FOI of about 5 per host per year in the pre-IRS season as estimated in Ghana (Figure 3) it seems that this would not translate to 4-year-old being immune naïve, and certainly this would not necessarily generalize well to a school-aged child population or an adult population. 

      The reviewer is indeed correct about the difficulty of empirically measuring the duration of infection for 1-5-year-olds, and that of further testing whether these 1-5-year-olds exhibit the same distribution for duration of infection as naïve adults co-infected with syphilis. We will nevertheless continue to use the described method for duration of infection, while better acknowledging and discussing the limitations this aspect of the method introduces. We note that the infection duration from the historical clinical data we have relied on, is being used in the malaria modeling community as one of the credible sources for this parameter of untreated natural infections in malaria-naïve individuals in malaria-endemic settings of Africa (e.g. in the agent-based model OpenMalaria, see 1).

      It is important to emphasize that the proposed methods apply to the MOI estimates for naïve or close to naïve patients. They are not suitable for FOI inference for the school-aged children and the adult populations of high-transmission endemic regions, since individuals in these age classes have been infected many times and their duration of infection is significantly shortened by their immunity. To reduce the degree of misspecification in infection duration and take full advantage of our proposed methods, we will emphasize in the revision the need to prioritize in future data collection and sampling efforts the subpopulation class who has received either no infection or a minimum number of infections in the past, and whose immune profile is close to that of naïve adults, for example, infants. This emphasis is aligned with the top priority of all intervention efforts in the short term, which is to monitor and protect the most vulnerable individuals from severe clinical symptoms and death.

      Also, force of infection for naïve hosts is a key basic parameter for epidemiological models of a complex infectious disease such as falciparum malaria, whether for agent-based formulations or equation-based ones. This is because force of infection for non-naïve hosts is typically a function of their immune status and the force of infection of naïve hosts. Thus, knowing the force of infection of naïve hosts can help parameterize and validate these models by reducing degrees of freedom.

      b. The evaluation of the capacity parameter c seems to be quite important and is set at 30, however, the authors only describe trying values of 25 and 30, and claim that this does not impact FOI inference, however it is not clear that this is the case. What happens if the carrying capacity is increased substantially? Alternatively, this would be more convincing if the authors provided a mathematical explanation of why the carrying capacity increase will not influence the FOI inference, but absent that, this should be mentioned and discussed as a limitation. 

      Thank you for this question. We will investigate more values of the parameter c systematically, including substantially higher ones. We note however that this quantity is the carrying capacity of the queuing system, or the maximum number of blood-stage strains that an individual human host can be co-infected with. We do have empirical evidence for the value of the latter being around 20 (2). This observed value provides a lower bound for parameter c. To account for potential under-sampling of strains, we thus tried values of 25 and 30 in the first version of our manuscript.

      In general, this parameter influences the steady-state queue length distribution based on the two-moment approximation, more specifically, the tail of this distribution when the flow of customers/infections is high. Smaller values of parameter c put a lower cap on the maximum value possible for the queue length distribution. The system is more easily “overflowed”, in which case customers (or infections) often find that there is no space available in the queuing system/individual host upon their arrival. These customers (or infections) will not increment the queue length. The parameter c has therefore a small impact for the part of the grid resulting in low flows of customers/infection, for which the system is unlikely to be overflowed. The empirical MOI distribution centers around 4 or 5 with most values well below 10, and only a small fraction of higher values between 15-20 (2). When one increases the value of c, the part of the grid generating very high flows of customers/infections results in queue length distributions with a heavy tail around large MOI values that are not supported by the empirical distribution. We therefore do not expect that substantially higher values for parameter c would change either the relative shape of the likelihood or the MLE.

      Reviewer #2 (Public Review):

      Summary:

      The authors combine a clever use of historical clinical data on infection duration in immunologically naive individuals and queuing theory to infer the force of infection (FOI) from measured multiplicity of infection (MOI) in a sparsely sampled setting. They conduct extensive simulations using agent-based modeling to recapitulate realistic population dynamics and successfully apply their method to recover FOI from measured MOI. They then go on to apply their method to real-world data from Ghana before and after an indoor residual spraying campaign.

      Strengths:

      (1) The use of historical clinical data is very clever in this context. 

      (2) The simulations are very sophisticated with respect to trying to capture realistic population dynamics. 

      (3) The mathematical approach is simple and elegant, and thus easy to understand. 

      Weaknesses: 

      (1) The assumptions of the approach are quite strong and should be made more clear. While the historical clinical data is a unique resource, it would be useful to see how misspecification of the duration of infection distribution would impact the estimates. 

      We thank the reviewer for bringing up the limitation of our proposed methods due to their reliance on a known and fixed duration of infection from historical clinical data. Please see our response to reviewer 1 comment 2a.

      (2) Seeing as how the assumption of the duration of infection distribution is drawn from historical data and not informed by the data on hand, it does not substantially expand beyond MOI. The authors could address this by suggesting avenues for more refined estimates of infection duration. 

      We thank the reviewer for pointing out a potential improvement to the work. We acknowledge that FOI is inferred from MOI, and thus is dependent on the information contained in MOI. FOI reflects risk of infection, is associated with risk of clinical episodes, and can relate local variation in malaria burden to transmission better than other proxy parameters for transmission intensity. It is possible that MOI can be as informative as FOI when one regresses the risk of clinical episodes and local variation in malaria burden with MOI. But MOI by definition is a number and not a rate parameter. FOI for naïve hosts is a key basic parameter for epidemiological models. This is because FOI of non-naïve hosts is typically a function of their immune status and the FOI of naïve hosts. Thus, knowing the FOI of naïve hosts can help parameterize and validate these models by reducing degrees of freedom. In this sense, we believe the transformation from MOI to FOI provides a useful step.

      Given the difficulty of measuring infection duration, estimating infection duration and FOI simultaneously appears to be an attractive alternative, as the referee pointed out. This will require however either cohort studies or more densely sampled cross-sectional surveys due to the heterogeneity in infection duration across a multiplicity of factors. These kinds of studies have not been, and will not be, widely available across geographical locations and time. This work aims to utilize more readily available data, in the form of sparsely sampled single-time-point cross-sectional surveys.

      (3) It is unclear in the example how their bootstrap imputation approach is accounting for measurement error due to antimalarial treatment. They supply two approaches. First, there is no effect on measurement, so the measured MOI is unaffected, which is likely false and I think the authors are in agreement. The second approach instead discards the measurement for malaria-treated individuals and imputes their MOI by drawing from the remaining distribution. This is an extremely strong assumption that the distribution of MOI of the treated is the same as the untreated, which seems unlikely simply out of treatment-seeking behavior. By imputing in this way, the authors will also deflate the variability of their estimates. 

      We thank the reviewer for pointing out aspects of the work that can be further clarified. It is difficult to disentangle the effect of drug treatment on measurement, including infection status, MOI, and duration of infection. Thus, we did not attempt to address this matter explicitly in the original version of our manuscript. Instead, we considered two extreme scenarios which bound reality, well summarized by the reviewer. First, if drug treatment has had no impact on measurement, the MOI of the drug-treated 1-5-year-olds would reflect their true underlying MOI. We can then use their MOI directly for FOI inference. Second, if the drug treatment had a significant impact on measurement, i.e., if it completely changed the infection status, MOI, and duration infection of drug-treated 1-5-year-olds, we would need to either exclude those individuals’ MOI or impute their true underlying MOI. We chose to do the latter in the original version of the manuscript. If those 1-5-year-olds had not received drug treatment, they would have had similar MOI values than those of the non-treated 1-5-year-olds. We can then impute their MOI by sampling from the MOI estimates of non-treated 1-5-year-olds.

      The reviewer is correct in pointing out that this imputation does not add additional information and can potentially deflate the variability of MOI distributions, compared to simply throwing or excluding those drug-treated 1-5-year-olds from the analysis. Thus, we can include in our revision FOI estimates with the drug-treated 1-5-year-olds excluded in the estimation.

      - For similar reasons, their imputation of microscopy-negative individuals is also questionable, as it also assumes the same distributions of MOI for microscopy-positive and negative individuals. 

      We imputed the MOI values of microscopy-negative but PCR-positive 1-5-year-olds by sampling from the microscopy-positive 1-5-year-olds, effectively assuming that both have the same, or similar, MOI distributions. We did so because there is a weak relationship in our Ghana data between the parasitemia level of individual hosts and their MOI (or detected number of var genes, on the basis of which the MOI values themselves were estimated). Parasitemia levels underlie the difference in detection sensitivity of PCR and microscopy.

      We will elaborate on this matter in our revised manuscript and include information from our previous and on-going work on the weak relationship between MOI/the number of var genes detected within an individual host and their parasitemia levels. We will also discuss potential reasons or hypotheses for this pattern.

      Reviewer #3 (Public Review):

      Summary: 

      It has been proposed that the FOI is a method of using parasite genetics to determine changes in transmission in areas with high asymptomatic infection. The manuscript attempts to use queuing theory to convert multiplicity of infection estimates (MOI) into estimates of the force of infection (FOI), which they define as the number of genetically distinct blood-stage strains. They look to validate the method by applying it to simulated results from a previously published agent-based model. They then apply these queuing theory methods to previously published and analysed genetic data from Ghana. They then compare their results to previous estimates of FOI. 

      Strengths: 

      It would be great to be able to infer FOI from cross-sectional surveys which are easier and cheaper than current FOI estimates which require longitudinal studies. This work proposes a method to convert MOI to FOI for cross-sectional studies. They attempt to validate this process using a previously published agent-based model which helps us understand the complexity of parasite population genetics. 

      Weaknesses: 

      (1) I fear that the work could be easily over-interpreted as no true validation was done, as no field estimates of FOI (I think considered true validation) were measured. The authors have developed a method of estimating FOI from MOI which makes a number of biological and structural assumptions. I would not call being able to recreate model results that were generated using a model that makes its own (probably similar) defined set of biological and structural assumptions a validation of what is going on in the field. The authors claim this at times (for example, Line 153 ) and I feel it would be appropriate to differentiate this in the discussion. 

      We thank the reviewer for this comment, although we think there is a mis-understanding on what can and cannot be practically validated in the sense of a “true” measure of FOI that would be free from assumptions for a complex disease such as malaria. We would not want the results to be over-interpreted and will extend the discussion of what we have done to test the methods. We note that for the performance evaluation of statistical methods, the use of simulation output is quite common and often a necessary and important step. In some cases, the simulation output is generated by dynamical models, whereas in others, by purely descriptive ones. All these models make their own assumptions which are necessarily a simplification of reality. The stochastic agent-based model (ABM) of malaria transmission utilized in this work has been shown to reproduce several important patterns observed in empirical data from high-transmission regions, including aspects of strain diversity which are not represented in simpler models.

      In what sense this ABM makes a set of biological and structural assumptions which are “probably similar” to those of the queuing methods we present, is not clear to us. We agree that relying on models whose structural assumptions differ from those of a given method or model to be tested, is the best approach. Our proposed methods for FOI inference based on queuing theory rely on the duration of infection distribution and the MOI distribution among sampled individuals, both of which can be direct outputs from the ABM. But these methods are agnostic on the specific mechanisms or biology underlying the regulation of duration and MOI.

      Another important point raised by this comment is what would be the “true” FOI value against which to validate our methods. Empirical MOI-FOI pairs for FOI measured directly by tracking cohort studies are still lacking. There are potential measurement errors for both MOI and FOI because the polymorphic markers typically used in different cohort studies cannot differentiate hyper-diverse antigenic strains fully and well (5). Also, these cohort studies usually start with drug treatment. Alternative approaches do not provide a measure of true FOI, in the sense of the estimation being free from assumptions. For example, one approach would be to fit epidemiological models to densely sampled/repeated cross-sectional surveys for FOI inference. In this case, no FOI is measured directly and further benchmarked against fitted FOI values. The evaluation of these models is typically based on how well they can capture other epidemiological quantities which are more easily sampled or measured, including prevalence or incidence. This is similar to what is done in this work. We selected the FOI values that maximize the likelihood of observing the given distribution of MOI estimates. Furthermore, we paired our estimated FOI value for the empirical data from Ghana with another independently measured quantity EIR (Entomological Inoculation Rate), typically used in the field as a measure of transmission intensity. We check whether the resulting FOI-EIR point is consistent with the existing set of FOI-EIR pairs and the relationship between these two quantities from previous studies. We acknowledge that as for model fitting approaches for FOI inference, our validation is also indirect for the field data.

      Prompted by the reviewer’s comment, we will discuss this matter in more detail in our revised manuscript, including clarifying further certain basic assumptions of our agent-based model, emphasizing the indirect nature of the validation with the field data and the existing constraints for such validation.

      (2) Another aspect of the paper is adding greater realism to the previous agent-based model, by including assumptions on missing data and under-sampling. This takes prominence in the figures and results section, but I would imagine is generally not as interesting to the less specialised reader. The apparent lack of impact of drug treatment on MOI is interesting and counterintuitive, though it is not really mentioned in the results or discussion sufficiently to allay my confusion. I would have been interested in understanding the relationship between MOI and FOI as generated by your queuing theory method and the model. It isn't clear to me why these more standard results are not presented, as I would imagine they are outputs of the model (though happy to stand corrected - it isn't entirely clear to me what the model is doing in this manuscript alone). 

      We thank the reviewer for this comment. We will add supplementary figures for the MOI distributions generated by the queuing theory method (i.e., the two-moment approximation method) and our agent-based model in our revised manuscript.

      In the first version of our manuscript, we considered two extreme scenarios which bound the reality, instead of simply assuming that drug treatment does not impact the infection status, MOI, and duration of infection. See our response to reviewer 2 point (3). The resulting FOI estimates differ but not substantially across the two extreme scenarios, partially because drug-treated individuals’ MOI distribution is similar to that of non-treated individuals (or the apparent lack of drug treatment on MOI as pointed by the referee). We will consider potentially adding some formal test to quantify the difference between the two MOI distributions and how significant the difference is. We will discuss which of the two extreme scenarios reality is closer to, given the result of the formal test. We will also discuss in our revision possible reasons/hypotheses underlying the impact of drug treatment on MOI from the perspective of the nature, efficiency, and duration of the drugs administrated.

      Regarding the last point of the reviewer, on understanding the relationship between MOI and FOI, we are not fully clear about what was meant. We are also confused about the statement on what the “model is doing in this manuscript alone”. We interpret the overall comment as the reviewer suggesting a better understanding of the relationship between MOI and FOI, either between their distributions, or the moments of their distributions, perhaps by fitting models including simple linear regression models. This approach is in principle possible, but it is not the focus of this work. It will be equally difficult to evaluate the performance of this alternative approach given the lack of MOI-FOI pairs from empirical settings with directly measured FOI values (from large cohort studies). Moreover, the qualitative relationship between the two quantities is intuitive. Higher FOI values should correspond to higher MOI values. Less variable FOI values should correspond to more narrow or concentrated MOI distributions, whereas more variable FOI values should correspond to more spread-out ones. We will discuss this matter in our revised manuscript.

      (3) I would suggest that outside of malaria geneticists, the force of infection is considered to be the entomological inoculation rate, not the number of genetically distinct blood-stage strains. I appreciate that FOI has been used to explain the latter before by others, though the authors could avoid confusion by stating this clearly throughout the manuscript. For example, the abstract says FOI is "the number of new infections acquired by an individual host over a given time interval" which suggests the former, please consider clarifying. 

      We thank the reviewer for this helpful comment as it is fundamental that there is no confusion on the basic definitions. EIR, the entomological inoculation rate, is closely related to the force of infection but is not equal to it. EIR focuses on the rate of arrival of infectious bites and is measured as such by focusing on the mosquito vectors that are infectious and arrive to bite a given host. Not all these bites result in actual infection of the human host. Epidemiological models of malaria transmission clearly make this distinction, as FOI is defined as the rate at which a host acquires infection. This definition comes from more general models for the population dynamics of infectious diseases in general. (For diseases simpler than malaria, with no super-infection, the typical SIR models define the force of infection as the rate at which a susceptible individual becomes infected).  For malaria, force of infection refers to the number of blood-stage new infections acquired by an individual host over a given time interval. This distinction between EIR and FOI is the reason why studies have investigated their relationship, with the nonlinearity of this relationship reflecting the complexity of the underlying biology and how host immunity influences the outcome of an infectious bite.

      We agree however with the referee that there could be some confusion in our definition resulting from the approach we use to estimate the MOI distribution (which provides the basis for estimating FOI). In particular, we rely on the non-existent to very low overlap of var repertoires among individuals with MOI=1, an empirical pattern we have documented extensively in previous work (See 2, 3, and 4). The method of var_coding and its Bayesian formulation rely on the assumption of negligible overlap. We note that other approaches for estimating MOI (and FOI) based on other polymorphic markers, also make this assumption (reviewed in _5). Ultimately, the FOI we seek to estimate is the one defined as specified above and in both the abstract and introduction, consistent with the epidemiological literature. We will include clarification in the introduction and discussion of this point in the revision.

      (4) Line 319 says "Nevertheless, overall, our paired EIR (directly measured by the entomological team in Ghana (Tiedje et al., 2022)) and FOI values are reasonably consistent with the data points from previous studies, suggesting the robustness of our proposed methods". I would agree that the results are consistent, given that there is huge variation in Figure 4 despite the transformed scales, but I would not say this suggests a robustness of the method. 

      We will modify the relevant sentences to use “consistent” instead of “robust”.

      (5) The text is a little difficult to follow at times and sometimes requires multiple reads to understand. Greater precision is needed with the language in a few situations and some of the assumptions made in the modelling process are not referenced, making it unclear whether it is a true representation of the biology. 

      We thank the reviewer for this comment. As also mentioned in the response to reviewer 1’s comments, we will reorganize and rewrite parts of the text in our revision to improve clarity.

      References and Notes

      (1) Maire, N. et al. A model for natural immunity to asexual blood stages of Plasmodium falciparum malaria in endemic areas. Am J Trop Med Hyg., 75(2 Suppl):19-31 (2006).

      (2) Tiedje, K. E. et al. Measuring changes in Plasmodium falciparum census population size in response to sequential malaria control interventions. eLife, 12 (2023).

      (3) Day, K. P. et al. Evidence of strain structure in Plasmodium falciparum var gene repertoires in children from Gabon, West Africa. Proc. Natl. Acad. Sci. U.S.A., 114(20), 4103-4111 (2017).

      (4) Ruybal-Pesántez, S. et al. Population genomics of virulence genes of Plasmodium falciparum in clinical isolates from Uganda. Sci. Rep., 7(11810) (2017).

      (5) Labbé, F. et al. Neutral vs. non-neutral genetic footprints of Plasmodium falciparum multiclonal infections. PLoS Comput Biol 19(1) (2023).

    2. Reviewer #1 (Public Review):

      Summary:

      In their paper, Zhan et al. have used Pf genetic data from simulated data and Ghanaian field samples to elucidate a relationship between multiplicity of infection (MOI) (the number of distinct parasite clones in a single host infection) and force of infection (FOI). Specifically, they use sequencing data from the var genes of Pf along with Bayesian modeling to estimate MOI individual infections and use these values along with methods from queueing theory that rely on various assumptions to estimate FOI. They compare these estimates to known FOIs in a simulated scenario and describe the relationship between these estimated FOI values and another commonly used metric of transmission EIR (entomological inoculation rate).

      This approach does fill an important gap in malaria epidemiology, namely estimating the force of infection, which is currently complicated by several factors including superinfection, unknown duration of infection, and highly genetically diverse parasite populations. The authors use a new approach borrowing from other fields of statistics and modeling and make extensive efforts to evaluate their approach under a range of realistic sampling scenarios. However, the write-up would greatly benefit from added clarity both in the description of methods and in the presentation of the results. Without these clarifications, rigorously evaluating whether the author's proposed method of estimating FOI is sound remains difficult. Additionally, there are several limitations that call into question the stated generalizability of this method that should at minimum be further discussed by authors and in some cases require a more thorough evaluation.

      Major comments:

      (1) Description and evaluation of FOI estimation procedure.

      a. The methods section describing the two-moment approximation and accompanying appendix is lacking several important details. Equations on lines 891 and 892 are only a small part of the equations in Choi et al. and do not adequately describe the procedure notably several quantities in those equations are never defined some of them are important to understand the method (e.g. A, S as the main random variables for inter-arrival times and service times, aR and bR which are the known time average quantities, and these also rely on the squared coefficient of variation of the random variable which is also never introduced in the paper). Without going back to the Choi paper to understand these quantities, and to understand the assumptions of this method it was not possible to follow how this works in the paper. At a minimum, all variables used in the equations should be clearly defined.

      b. Additionally, the description in the main text of how the queueing procedure can be used to describe malaria infections would benefit from a diagram currently as written it's very difficult to follow.

      c. Just observing the box plots of mean and 95% CI on a plot with the FOI estimate (Figures 1, 2, and 10-14) is not sufficient to adequately assess the performance of this estimator. First, it is not clear whether the authors are displaying the bootstrapped 95%CIs or whether they are just showing the distribution of the mean FOI taken over multiple simulations, and then it seems that they are also estimating mean FOI per host on an annual basis. Showing a distribution of those per-host estimates would also be helpful. Second, a more quantitative assessment of the ability of the estimator to recover the truth across simulations (e.g. proportion of simulations where the truth is captured in the 95% CI or something like this) is important in many cases it seems that the estimator is always underestimating the true FOI and may not even contain the true value in the FOI distribution (e.g. Figure 10, Figure 1 under the mid-IRS panel). But it's not possible to conclude one way or the other based on this visualization. This is a major issue since it calls into question whether there is in fact data to support that these methods give good and consistent FOI estimates.

      d. Furthermore the authors state in the methods that the choice of mean and variance (and thus second moment) parameters for inter-arrival times are varied widely, however, it's not clear what those ranges are there needs to be a clear table or figure caption showing what combinations of values were tested and which results are produced from them, this is an essential component of the method and it's impossible to fully evaluate its performance without this information. This relates to the issue of selecting the mean and variance values that maximize the likelihood of observing a given distribution of MOI estimates, this is very unclear since no likelihoods have been written down in the methods section of the main text, which likelihood are the authors referring to, is this the probability distribution of the steady state queue length distribution? At other places the authors refer to these quantities as Maximum Likelihood estimators, how do they know they have found the MLE? There are no derivations in the manuscript to support this. The authors should specify the likelihood and include in an appendix an explanation of why their estimation procedure is in fact maximizing this likelihood, preferably with evidence of the shape of the likelihood, and how fine the grid of values they tested is for their mean and variance since this could influence the overall quality of the estimation procedure.

      (2) Limitation of FOI estimation procedure.

      a. The authors discuss the importance of the duration of infection to this problem. While I agree that empirically estimating this is not possible, there are other options besides assuming that all 1-5-year-olds have the same duration of infection distribution as naïve adults co-infected with syphilis. E.g. it would be useful to test a wide range of assumed infection duration and assess their impact on the estimation procedure. Furthermore, if the authors are going to stick to the described method for duration of infection, the potentially limited generalizability of this method needs to be further highlighted in both the introduction, and the discussion. In particular, for an estimated mean FOI of about 5 per host per year in the pre-IRS season as estimated in Ghana (Figure 3) it seems that this would not translate to 4-year-old being immune naïve, and certainly this would not necessarily generalize well to a school-aged child population or an adult population.

      b. The evaluation of the capacity parameter c seems to be quite important and is set at 30, however, the authors only describe trying values of 25 and 30, and claim that this does not impact FOI inference, however it is not clear that this is the case. What happens if the carrying capacity is increased substantially? Alternatively, this would be more convincing if the authors provided a mathematical explanation of why the carrying capacity increase will not influence the FOI inference, but absent that, this should be mentioned and discussed as a limitation.

    1. Gloria Mark, a professor of information science at the University of California, Irvine, and the author of “Attention Span,” started researching the way people used computers in 2004. The average time people spent on a single screen was 2.5 minutes. “I was astounded,” she told me. “That was so much worse than I’d thought it would be.” But that was just the beginning. By 2012, Mark and her colleagues found the average time on a single task was 75 seconds. Now it’s down to about 47.
    1. Writing is not merely a mode of communication. It’s a process that, if we move beyond simple formulas, forces us to reflect, think, analyze and reason

      This statement stood out to me because it states how writing is more than just a form of expression. Its more complex meaning it forces us to analyze different things when we're writing.

    1. The first draft is the child’s draft, where you let it all pour out and then let it romp all over the place, knowing that no one is going to see it and that you can shape it later. You just let this childlike part of you channel whatever voices and visions come through and onto the page.

      i agree with this because personally most of the time i want my first draft is perfect but in all honesty it's called a draft for a reason it's so when can just throw out ideas and fix it up later.

    1. Welcome back.

      This demo is going to bring together some really important theory and architecture that you've learned over the past few lessons.

      What we're starting this demo lesson with is this architecture.

      We have our VPC, the Animals for Life VPC in US East 1.

      It uses the 10.16.0.0/16 side range.

      It has 12 subnets created inside it, over three AZs with four tiers, Reserve, DB, Application and Web.

      Now currently all the subnets are private and can't be used for communication with the internet or the AWS public zone.

      In this demo we want to reconfigure the VPC to allow that.

      So the first step is to create an internet gateway and attach it.

      To do that, I'm going to move across to my desktop.

      Now to do this in your environment, you'll need the VPC and subnet configuration as you set it up in the previous demo lesson.

      So that configuration needs to be in place already.

      You need to be logged in as the I am admin user of the management account of the organization and have the Northern Virginia region selected, so US - East - 1.

      So go ahead and move across to the VPC console.

      Now this should already be in the recently visited services because you were using this in the previous demo lesson, but if it's not visible just click in the services drop down, type VPC and then click to move to the VPC console.

      Now if you do still have the configuration as it was at the end of the previous demo lesson, you should be able to click on subnets on the menu on the left and see a list of lots of subnets.

      You'll see the ones for the default VPC without a name.

      And if you have the correct configuration, you should see a collection of 12 subnets, 3 application subnets, 3 database subnets, 3 reserved subnets and then 3 web subnets.

      So all of these should be in place within the Animals for Life VPC in order to do the tasks within this demo lesson.

      So I'm going to assume from this point onwards that you do have all of these subnets created and configured.

      Now what we're going to be doing in this demo lesson is configuring the 3 web subnets, so web A, web B and web C, to be public subnets.

      Being a public subnet means that you can launch resources into the subnet, have them allocated with a public IP version 4 address and have connectivity to and from the public IP version 4 internet.

      And in order to enable that functionality, there are a number of steps that we need to perform and I want you to get the practical experience of implementing these within your own environment.

      Now the first step to making subnets public is that we need an internet gateway attached to this VPC.

      So internet gateways, as I talked about in the previous theory lesson, are highly available gateway objects which can be used to allow public routing to and from the internet.

      So we need to create one.

      So let's click on internet gateways on the menu on the left.

      They'll already be an internet gateway in place for the default VPC.

      Remember when you created the default VPC, all this networking infrastructure is created and configured on your behalf.

      But because we created a custom VPC for animals for life, we need to do this manually.

      So to do that, go ahead and click on create internet gateway.

      We're going to call the internet gateway A4L, so animals for life, VPC 1, which is the VPC we're going to attach it to and then IGW for internet gateway.

      So A4L-VPC1-IGW.

      Now that's the only information that we need to enter, so scroll down and click on create internet gateway.

      Internet gateways are initially not attached to a VPC and we can tell that because it's initially in.

      We need to attach this to the animals for life VPC.

      So click on actions and then attach to VPC inside the available VPCs box.

      Just click and then select A4L-IVAN-VPC1.

      Once selected, go ahead and click on attach internet gateway and that will attach our brand new internet gateway to the animals for life VPC.

      And that means that it's now available within that VPC as a gateway object, which gives the VPC the capability to communicate to and receive communications from the public internet and the AWS public zone.

      Now the next step is that we want to make all the subnets in the web tier public, so the services deployed into these subnets can take advantage of this functionality.

      So we want the web subnets to be able to communicate to and receive communications from the public internet and AWS public services.

      Now there are a number of steps that we need to do to accomplish this.

      We need to create a route table for the public subnets.

      We need to associate this route table with the three public subnets, so web A, web B and web C and then we need to add two routes to this route table.

      One route will be a default route for IP version 4 traffic and the other will be a default route for IP version 6 traffic.

      And both of these routes for their target will be pointing at the internet gateway that you've just created and attached to this VPC.

      Now this will configure the VPC router to forward any data intended for anything not within our VPC to the internet gateway.

      Finally, on each of these web subnets will be configuring the subnet to auto assign a public IP version 4 address and that will complete the process of making them public.

      So let's perform all of these sets of configuration.

      So now that we're back at the AWS console, we need to create a route table.

      So go ahead and click on route tables on the menu on the left and then we're going to create a new route table.

      First we'll select the VPC that this route table will belong to and it's going to be the animals for life.

      If a VPC, so go ahead and select that VPC and then we're going to give this route table a name.

      And I like to keep the naming scheme consistent, so we're going to use A4, L4, animals for life and then a hyphen.

      VPC1 because this is the VPC the route table will belong to and then hyphen RT for route table and then hyphen and then web because this route table is going to be used for the web subnets.

      So go ahead and create this route table and click route tables on the menu on the left.

      If we select the route table that we've just created, so that's the one that's called A4, L hyphen VPC1 hyphen RT hyphen web and then just expand this overview area at the bottom.

      We'll be able to see all the information about this route table.

      Now there are a number of areas of this which are important to understand.

      One is the routes area which lists all the routes on this route table and the other is subnet associations.

      This determines which subnets this route table is.

      So let's go to subnet associations and currently we can see that it's not actually associated with any subnets within this VPC.

      We need to adjust that so go ahead and edit those associations and we're going to associate it with the three web subnets.

      So you need to select web A, web B and web C.

      Now notice how all those are currently associated with the main route table of the VPC.

      Remember a subnet can only be associated with one route table at a time.

      If you don't explicitly associate a route table with a subnet then it's associated with the main route table.

      We're going to change that.

      We're going to explicitly associate this new route table with the web A, web B and web C subnets.

      So go ahead and say that.

      So now this route table has been associated with web A, web B and web C.

      Those subnets are no longer associated with the main route table of the VPC.

      So now we've configured the association as a route to routes.

      And we can see that this route table has two local routes.

      We've got the IP version 4 side of the VPC and the app.

      IP version 6 side of the VPC.

      So these two routes on this route table will mean that web A, web B and web C will know how to direct traffic towards any other IP version 4 or IP version 6 addresses within this VPC.

      Now these local routes can never be adjusted or removed, but what we can do is add additional routes.

      So we're going to add two routes, a default route for IP version 4 and a default route for IP version 6.

      So we'll do that.

      We'll start with IP version 4.

      So we'll edit those routes and then we'll add a route.

      The format for the IP version 4 default route is 0.0.0.0/0.

      And this means any IP addresses.

      Now I've talked elsewhere in the course how there is a priority to routing.

      Within a VPC there's a more specific route always takes priority.

      So this route, the /16 is more specific than this default route.

      So this default route will only affect IP version 4 traffic, which is not matched by this local route.

      So essentially anything which is IP version 4, which is not destined for the VPC, will use this default route.

      Now we need to pick the internet gateway as the target for this route.

      So click in the target box on this row, select internet gateway.

      There should only be one that's highlighted.

      That's the Animals for Life internet gateway you created moments ago.

      So select that and that means that any IP version 4 traffic which is not destined for the VPC side of it. [siren] Range will be sent to the Internet Gateway.

      Now we're going to do the same for IPv6.

      So go ahead and add another route.

      And the format for IPv60 default routes is double colon, forward slash zero.

      And this is the same architecture. [siren] It essentially means this matches all IPv6 addresses, but it's less specific than the IPv6 and version 6 local route on this top row.

      So this will only be used for any IPv6 addresses which are not in the IPv6 VPC side range.

      So go ahead and select Target, go to Internet Gateway, and select the Animals for Life Internet Gateway.

      And once you've done both of those, go ahead and click on Save Changes.

      Now this means that we now have two default routes, an IPv4 default route, and an IPv6 default route.

      So this means that anything which is associated with these route tables will now send any unknown traffic towards the Internet Gateway.

      But what we need to do before this works is we need to ensure that any resources launched into the Web A, Web B, or Web C subnets are allocated with public IPv4 addresses.

      To do that, go ahead and click on Subnets.

      In the list, we need to locate Web A, Web B, and Web C.

      So we'll start with Web A, so select Web A, click on Actions, and then Edit Subnet Settings.

      And this time, we're going to modify this subnet so that it automatically assigns a public IPv4 address.

      So check this box into the Save, and that means that any resources launched into the Web A subnet will be allocated with a public IPv4 address.

      Now we need to follow the same process for the other web subnets, so select the Web B subnet, click on Actions, and then Edit Subnet Settings.

      Enable IPv4, click on Save, and then do that same process for Web C.

      So locate Web C, click on Actions, and then Edit Subnet Settings.

      And then enable public IPv4 addresses and click on Save.

      So that's all the network configuration done.

      We've created an Internet Gateway.

      We've associated the Internet Gateway with the Animals for IPPC.

      We've created a Routetable for the web subnets.

      We've associated this Routetable with the web subnets.

      We've added default routes onto this Routetable, pointing at the Internet Gateway as a default IPv4 and IPv6 route.

      And then we've enabled the allocation of public IPv4 addresses for Web A, Web B, and Web C.

      Okay, so this is the end of Part 1 of this lesson.

      It was getting a little bit on the long side, and so I wanted to add a break.

      It's an opportunity just to take a rest or grab a coffee.

      Part 2 will be continuing immediately from the end of Part 1.

      To go ahead, complete the video, I'm ready, join me in part two.

    1. language is a natural growth and not an instrument which we shape for our own purposes.

      I think saying that the english language is in a decline doesn't make sense for this reason; maybe it's not a decline, perhaps just the slow movement of our language transitioning naturally and we shouldn't think of it as a collapse.

    1. Enslavers database lists nearly 60,000 individuals or commercial companies involved in the enslavement of Africans or people of African descent, lists the number of captives they enslaved, and provides links to the slave voyages in which they participated.

      I think it's really interesting how there's data on the enslavers and not just the Africans. Usually we don't see that side of history

    1. Whenever I cook, I find myself working just as she would, readying the ingredients—a mash of garlic, a julienne of red peppers, fantails of shrimp—and piling them in little mounds about the cutting surface. My mother never left me any recipes, but this is how I learned to make her food, each dish coming not from a list or a card but from the aromatic spread of a board.

      Sometimes the things that our parents do, we tend to follow those exact ways. From personal experience, my parents always use a plastic glove to throw away the garbage bags, so I feel it is necessary to use gloves everytime it’s my turn to throw away the garbage.

    2. My mother would gently set herself down in her customary chair near the stove. I sat across from her, my father and sister to my left and right, and crammed in the center was all the food I had made—a spicy codfish stew, say, or a casserole of gingery beef, dishes that in my youth she had prepared for us a hundred times.

      It's really important to recognize just how significant it means to pass down dishes from generation to generation because sometimes foods and entire cultural meals can be forgotten because the next generation was never taught

    1. leave “structure, language use and grammar” to AI to score while teachers look for “novelty, creativity and depth of insight.”

      I still am not comfortable suggesting that we can even leave this to AI. Just because it can generate proper syntax and readable prose, it doesn't mean it's qualified to EXPLAIN it because it still doesn't understand it.

    1. Note: This response was posted by the corresponding author to Review Commons. The content has not been altered except for formatting.

      Learn more at Review Commons


      Reply to the reviewers

      1. Point-by-point description of the revisions

      This section is mandatory. *Please insert a point-by-point reply describing the revisions that were already carried out and included in the transferred manuscript. *

      Reviewer #1 (Evidence, reproducibility, and clarity (Required)): This is an interesting manuscript from the Jagannathan laboratory, which addresses the interaction proteome of two satellite DNA-binding proteins, D1 and Prod. To prevent a bias by different antibody affinities they use GFP-fusion proteins of D1 and Prod as baits and purified them using anti GFP nanobodies. They performed the purifications in three different tissues: embryo, ovary and GSC enriched testes. Across all experiments, they identified 500 proteins with surprisingly little overlap among tissues and between the two different baits. Based on the observed interaction of prod and D1 with members of the canonical piRNA pathway the authors hypothesized that both proteins might influence the expression of transposable elements. However, neither by specific RNAi alleles or mutants that lead to a down regulation of D1 and Prod in the gonadal soma nor in the germline did they find an effect on the repression of transposable elements. They also did not detect an effect of a removal of piRNA pathway proteins on satellite DNA clustering, which is mediated by Prod and D1. However, they do observe a mis-localisation of the piRNA biogenesis complex to an expanded satellite DNA in absence of D1, which presumably is the cause of a mis-regulation of transposable elements in the F2 generation.This is an interesting finding linking satellite DNA and transposable element regulation in the germline. However, I find the title profoundly misleading as the link between satellite DNA organization and transgenerational transposon repression in Drosophila has not been identified by multi-tissue proteomics but by a finding of the Brennecke lab that the piRNA biogenesis complex has a tendency to localise to satellite DNA when the localisation to the piRNA locus is impaired. Nevertheless, the investigation of the D1 and Prod interactome is interesting and might reveal insights into the pathways that drive the formation of centromeres in a tissue specific manner.

      We thank the reviewer for the overall positive comments on our manuscript. As noted above, we have performed a substantial number of revision experiments and improved our text. We believe that our revised manuscript demonstrates a clear link between our proteomics data and the transposon repression. We would like to make three main points,

      1. Our proteomics data identified that D1 and Prod co-purified transposon repression proteins in embryos. To test the functional significance of this association, we have used Drosophila genetics to generate flies lacking embryonic D1. In adult ovaries from these flies, we observe a striking elevation in transposon expression and Chk2-dependent gonadal atrophy. Along with the results from the control genotypes (F1 D1 mutant, F2 D1 het), our data clearly indicate that embryogenesis (and potentially early larval development) are a period when D1 establishes heritable TE silencing that can persist throughout development.
      2. Based on the newly acquired RNA-seq and small RNA seq data, we have edited our title to more accurately reflect our data. Specifically, we have substituted the word 'transgenerational' with 'heritable', meaning that the presence of D1 during early development alone is sufficient to heritably repress TEs at later stages of development.
      3. In addition, our RNA seq and small RNA seq experiments demonstrate that D1 makes a negligible contribution to piRNA biogenesis and TE repression in adults, despite the significant mislocalization of the RDC complex. In this regard, our results are substantially different from the recent Kipferl study from the Brennecke lab (Baumgartner et al. 2022).

        Major comments Unfortunately, the proteomic data sets are not very convincing. Not even the corresponding baits are detected in all assays. I wonder whether the extraction procedure really allows the authors to analyze all functionally relevant interactions of Prod and D1. It would be good to see a western blot or an MS analysis of the soluble nuclear extract they use for purification compared to the insoluble chromatin. It may well be that a large portion of Prod or D1 is lost in this early step. I also find the description of the proteomic results hard to follow as the authors mostly list which proteins the find as interactors of Prod and D1 without stating in which tissue or with what bait they could purify them (i.e. p7: Importantly, our hits included known chromocenter-associated or pericentromeric heterochromatin-associated proteins, such as Su(var)3-9[52], ADD1[23,24], HP5[23,24],mei-S332[53], Mes-4[23], Hmr[24,39,54], Lhr[24,39], and members of the chromosome passenger complex, such as borr and Incenp[55]). It would be interesting to at least discuss tissue specific interactions.

      Out of six total AP-MS experiments in this manuscript (D1 x 3, Prod x 2 and Piwi), we observe a strong enrichment of the bait in 5/6 attempts (log2FC between 4-12). In the initial submission, the lack of a third high-quality biological replicate for the D1 testis sample meant that only the p-value (0.07) was not meeting the cutoff. To address this, we have repeated this experiment with an additional biological replicate (Fig. S2A), and our data now clearly show that D1 is significantly enriched in the testis sample.

      As suggested by the reviewer, we have also assessed our lysis conditions (450mM NaCl and benzonase) and the solubilization of our baits post-lysis. In Fig. S1D, we have blotted equivalent fractions of the soluble supernatant and insoluble pellet from GFP-Piwi embryos and show that both GFP-Piwi and D1 are largely solubilized following lysis. In Fig. S1E, we also show that our IP protocol works efficiently.

      GFP-Prod pulldown in embryos is the only instance in which we do not detect the bait by mass spectrometry. Here, one reason could be relatively low expression of GFP-Prod in comparison to GFP-D1 (Fig. S1E), which may lead to technical difficulties in detecting peptides corresponding to Prod. However, we note that Prod IP co-purified proteins such as Bocks that were previously suggested as Prod interactors from high-throughput studies (Giot et al. 2003; Guruharsha et al. 2011). In addition, Prod IP from embryos also co-purified proteins known to associate with chromocenters such as Hcs and Saf-B. Finally, the concordance between D1 and Prod co-purified proteins from embryo lysates (Fig. 2A, C) suggest that the Prod IP from embryos is of reasonable quality.

      We also acknowledge the reviewer's comment that the description of the proteomic data was hard to follow. Therefore, we have revised our text to clearly indicate which bait pulled down which protein in which tissue (lines 148-156). We have also highlighted and discussed bait-specific and tissue-specific interactions in the text (lines 162-173).

      Minor comments The authors may also want to provide a bit more information on the quantitation of the proteomic data such as how many values were derive from the match-between runs function and how many were imputed as this can severely distort the quantification.

      Figure 1: Distribution of data after imputation in embryo (left), ovary (middle) and testis (right) datasets. Imputation is performed with random sampling from the 1% least intense values generated by a normal distribution.

      To ensure the robustness of our data analysis, we considered only those proteins that were consistently identified in all replicates for at least one bait (GFP-D1, GFP-Prod or NLS-GFP). This approach resulted in a relative low number of missing values. However, it is also important to bear in mind that in an AP-MS experiment, the number of missing values is higher, as interactors are not identified in the control pulldown. Importantly, the imputation of missing values during the data analysis did not alter the normal distribution of the dataset (Fig. 1, this document). Detailed information regarding the imputed values is also provided (Table 1, this document). The coding script used for the data analysis is available in the PRIDE submission of the dataset (Table 2, this document). This information has been added to our methods section under data availability.

      Table 1: ____Number of match-between-runs and imputations for embryo, ovary and testis datasets

      Dataset

      #match-between-runs

      %match-between-runs

      %imputation

      Embryo

      5541/27543

      20.11%

      8.36%

      Ovary

      1936/9530

      20.30%

      8.18%

      Testis

      1748/7168

      24.39%

      3.12%

      Table 2: ____Access to the PRIDE submission of the datasets

      Name

      ID PRIDE

      UN reviewer

      PW reviewer

      IP-MS of D1 from Testis tissue

      PXD044026

      reviewer_pxd044026@ebi.ac.uk

      ydswDQVW

      IP-MS of Piwi from Embryo tissue

      PXD043237

      reviewer_pxd043237@ebi.ac.uk

      TMCoDsdx

      IP-MS of Prod and D1 proteins from Ovary tissue

      PXD043236

      reviewer_pxd043236@ebi.ac.uk

      VOHqPmaS

      IP-MS of Prod and D1 proteins from Embryo tissue

      PXD043234

      reviewer_pxd043234@ebi.ac.uk

      L77VXdvA

      **Referee Cross-Commenting** I agree with the two other reviewers that the connection between the interactome and the transgenerational phenotype is unclear. This is also what I meant i my comment that the title is somewhat misleading. A systematic analysis of the D1 and Prod knock down effect on mRNAs and small Rnas would indeed be helpful to better understand the interesting effect.

      As suggested by the reviewer, we have performed RNA seq and small RNA seq in control and D1 mutant ovaries (Fig. 4) to fully understand the contribution of D1 in piRNA biogenesis and TE repression. Briefly, the mislocalization of RDC complex in D1 mutant ovaries does not significantly affect TE-mapping piRNA biogenesis (Fig. 4C, E). In addition, loss of D1 does not substantially alter TE expression in the ovaries (Fig. 4B) or alter the expression of genes involved in TE repression (Fig. 4F). Along with the results presented in Fig. 5 and Fig. 6, our data clearly indicate that embryogenesis (and potentially early larval development) is a critical period during which D1 makes an important contribution to TE repression.

      Reviewer #1 (Significance (Required)): Nevertheless, the investigation of the D1 and Prod interactome is interesting and might reveal insights into the pathways that drive the formation of centromeres in a tissue specific manner. It may be mostly interesting for the Drosophila community but could also be exiting for a broader audience interested in the connection of heterochromatin and its indirect effect on the regulation of transposable elements.

      We thank the reviewer again for the helpful and constructive comments, which have enabled us to significantly improve our study. We are excited by the results from our study, which illuminate unappreciated aspects of transcriptional silencing in constitutive heterochromatin.

      Reviewer #2 (Evidence, reproducibility and clarity (Required)): Chavan et al. set out to enrich our compendium of pericentric heterochromatin-associated proteins - and to learn some new biology along the way. An ambitious AP-Mass baited with two DNA satellite-binding proteins (D1 and Prod), conducted across embryos, ovaries, and testes, yielded hundreds of candidate proteins putatively engaged at chromocenters. These proteins are enriched for a restricted number of biological pathways, including DNA repair and transposon regulation. To investigate the latter in greater depth, the authors examine D1 and prod mutants for transposon activity changes using reporter constructs for multiple elements. These reporter constructs revealed no transposon activation in the adult ovary, where many proteins identified in the mass spec experiments restrict transposons. However, the authors suggest that the D1 mutant ovaries do show disrupted localization of a key member of a transposon restriction pathway (Cuff), and infer that this mislocalization triggers a striking, transposon derepression phenotype in the F2 ovaries.

      The dataset produced by the AP-Mass Spec offers chromosome biologists an unprecedented resource. The PCH is long-ignored chromosomal region that has historically received minimal attention; consequently, the pathways that regulate heterochromatin are understudied. Moreover, attempting to connect genome organization to transposon regulation is a new and fascinating area. I can easily envision this manuscript triggering a flurry of discovery; however, there is quite a lot of work to do before the data can fully support the claims.

      We appreciate the reviewer taking the time to provide thoughtful comments and constructive suggestions to improve the manuscript. We believe that we have addressed all the comments raised to the significant advantage of our paper.

      Major comments 1. The introduction requires quite a radical restructure to better highlight the A) importance of the work and B) limit information whose relevance is not clear early in the manuscript. A. Delineating who makes up heterochromatin is a long-standing problem in chromosome biology. This paper has huge potential to contribute to this field; however, it is not the first. Others are working on this problem in other systems, for example PMID:29272703. Moreover, we have some understanding of the distinct pathways that may impact heterochromatin in different tissues (e.g., piRNA biology in ovaries vs the soma). Also, the mutant phenotypes of prod and D1 are different. Fleshing out these three distinct points could help the reader understand what we know and what we don't know about heterochromatin composition and its special biology. Understanding where we are as a field will offer clear predictions about who the interactors might be that we expect to find. For example, given the dramatically different D1 and prod mutant phenotypes (and allele swap phenotypes), how might the interactors with these proteins differ? What do we know about heterochromatin formation differences in different tissues? And how might these differences impact heterochromatin composition?

      The reviewer brings up a fair point and we have significantly reworked our introduction. We share the reviewer's opinion that improved knowledge of the constitutive heterochromatin proteome will reveal novel biology.

      1. The attempt to offer background on the piRNA pathway and hybrid dysgenesis in the Introduction does not work. As a naïve reader, it was not clear why I was reading about these pathways - it is only explicable once the reader gets to the final third of the Results. Moreover, the reader will not retain this information until the TE results are presented many pages later. I strongly urge the authors to shunt the two TE restriction paragraphs to later in the manuscript. They are currently a major impediment to understanding the power of the experiment - which is to identify new proteins, pathways, and ultimately, biology that are currently obscure because we have so little handle on who makes up heterochromatin.

      We agree with this suggestion. We have introduced the piRNA pathway in the results section (lines 205 - 216), right before this information is needed. We've also removed the details on hybrid dysgenesis, since our new data argues against a maternal effect from the D1 mutant.

      The implications of the failure to rescue female fertility by the tagged versions of both D1 and Prod are not discussed. Consequently, the reader is left uneasy about how to interpret the data.

      We understand this point raised by the reviewer. However, in our proteomics experiments, we have used GFP-D1 and GFP-Prod ovaries from ~1 day old females (line 579, methods). These ovaries are morphologically similar to the wild type (Fig. S1C) and their early germ cell development appears to be intact. Moreover, chromocenter formation in female GSCs is comparable to the wildtype (Fig. 1C-D). These data, along with the rescue of the viability of the Prod mutant (Fig. S1A-B), suggest that the presence of a GFP tag is not compromising D1 or Prod function in the early stages of germline development and is consistent with published and unpublished data from our lab. In addition, D1 and Prod from ovaries co-purify proteins such as Rfc38 (D1), Smn (D1), CG15107 (Prod), which have been identified in previous high-throughput screens (Guruharsha et al. 2011; Tang et al. 2023). For these reasons, we believe that GFP-D1 and GFP-Prod ovaries are a good starting point for the AP-MS experiment. We speculate that the failure to completely rescue female fertility may be due to improper expression levels of GFP-D1 or GFP-Prod flies at later stages of oogenesis, which are not present in ovaries from newly eclosed females and therefore unlikely to affect our proteomic data.

      1. How were the significance cut-offs determined? Is the p-value reported the adjusted p-value? As a non-expert in AP-MS, I was surprised to find that the p-value, at least according to the Methods, was not adjusted based on the number of tests. This is particularly relevant given the large/unwieldy(?) number of proteins that were identified as signficant in this study. Moreover, the D1 hit in Piwi pull down is actually not significant according to their criteria of p We used a standard cutoff of log2FC>1 and p2FC and low p-value) since these are more likely to be bona fide interactors. Third, we have used string-DB for functional analyses where we can place our hits in existing protein-protein interaction networks. Using this approach, we detect that Prod (but not D1) pulls down multiple members of the RPA complex in the embryo (RPA2 and RpA-70, Fig. S2B) while embryonic D1 (but not Prod) pulls down multiple components of the origin recognition complex (Orc1, lat, Orc5, Orc6, Fig. S2C) and the condensin I complex (Cap-G, Cap-D2, barr, Fig. S2D). Altogether, these filtering strategies allow us to eliminate as many false positives as possible while making sure to minimize the loss of true hits through multiple testing correction.

      How do we know if the lack of overlap across tissues is indeed germline- or soma-specialization rather than noise?

      To address this part of the comment, we have amended our text (lines 162-173) as follows,

      'We also observed a substantial overlap between D1- and Prod-associated proteins (yellow points in Fig. 2A, B, Table S1-3), with 61 hits pulled down by both baits (blue arrowheads, Fig. 2C) in embryos and ovaries. This observation is consistent with the fact that both D1 and Prod occupy sub-domains within the larger constitutive heterochromatin domain in nuclei. Surprisingly, only 12 proteins were co-purified by the same bait (D1 or Prod) across different tissues (magenta arrowheads, Fig. 2C, Table S1-3). In addition, only a few proteins such as an uncharacterized DnaJ-like chaperone, CG5504, were associated with both D1 and Prod in embryos and ovaries (Fig. 2D). One interpretation of these results is that the protein composition of chromocenters may be tailored to cell- and tissue-specific functions in Drosophila. However, we also note that the large number of unidentified peptides in AP-MS experiments means that more targeted experiments are required to validate whether certain proteins are indeed tissue-specific interactors of D1 and Prod.'

      To make this inference, conducting some validation would be required. More generally, I was surprised to see no single interactor validated by reciprocal IP-Westerns to validate the Mass-Spec results, though I am admittedly only adjacent to this technique. Note that colocalization, to my mind, does not validate the AP-MS data - in fact, we would a priori predict that piRNA pathway members would co-localize with PCH given the enrichment of piRNA clusters there.

      Here, we would point out that we have conducted multiple validation experiments with a specific focus on the functional significance of the associations between D1/Prod and TE repression proteins in embryos. While the reviewer suggests that piRNA pathway proteins may be expected to enrich at the pericentromeric heterochromatin, this is not always the case. For example, Piwi and Mael are present across the nucleus (pulled down by D1/Prod in embryos) while Cutoff, which is present adjacent to chromocenters in nurse cells, was not observed to interact with either D1 or Prod in the ovary samples.

      Therefore, for Piwi, we performed a reciprocal AP-MS experiment in embryos due to the higher sensitivity of this method (GFP-Piwi AP-MS, Fig. 3B). Excitingly, this experiment revealed that four largely uncharacterized proteins (CG14715, CG10208, Ugt35D1 and Fit) were highly enriched in the D1, Prod and Piwi pulldowns and these proteins will be an interesting cohort for future studies on transposon repression in Drosophila (Fig. 3C).

      Furthermore, we believe that determining the localization of proteins co-purified by D1/Prod is an important and orthogonal validation approach. For Sov, which is implicated in piRNA-dependent heterochromatin formation, we observe foci that are in close proximity to D1- and Prod-containing chromocenters (Fig. 3A).

      As for suggestion to validate by IP-WBs, we would point out that chromocenters exhibit properties associated with phase separated biomolecular condensates. Based on the literature, these condensates tend to associate with other proteins/condensates through low affinity or transient interactions that are rarely preserved in IP-WBs, even if there are strong functional relationships. One example is the association between D1 and Prod, which do not pull each other down in an IP-WB (Jagannathan et al. 2019), even though D1 and Prod foci dynamically associate in the nucleus and mutually regulate each other's ability to cluster satellite DNA repeats (Jagannathan et al. 2019). Similarly, IP-WB using GFP-Piwi embryos did not reveal an interaction with D1 (Fig. S4B). However, our extensive functional validations (Figures 4-6) have revealed a critical role for D1 in heritable TE repression.

      The AlphaFold2 data are very interesting but seem to lack of negative control. Is it possible to incorporate a dataset of proteins that are not predicted to interact physically to elevate the impact of the ones that you have focused on? Moreover, the structural modeling might suggest a competitive interaction between D1 and piRNAs for Piwi. Is this true? And even if not, how does the structural model contribute to your understanding for how D1 engages with the piRNA pathway? The Cuff mislocalization?

      In the revised manuscript, we have generated more structural models using AlphaFold Multimer (AFM) for proteins (log2FC>2, p0.5 and ipTM>0.8), now elaborated in lines 175-177. Despite the extensive disorder in D1 and Prod, we identified 22 proteins, including Piwi, that yield interfaces with ipTM scores >0.5 (Table S4, Table S8). These hits are promising candidates to further understand D1 and Prod function in the future.

      For the predicted model between Prod/D1 and Piwi (Fig. S4A), one conclusion could indeed be competition between D1/Prod and piRNAs for Piwi. Another possibility is that a transient interaction mediated by disordered regions on D1/Prod could recruit Piwi to satellite DNA-embedded TE loci in the pericentromeric heterochromatin. These types of interactions may be especially important in the early embryonic cycles, where repressive histone modifications such as H3K9me2/3 must be deposited at the correct loci for the first time. We suggest that mutating the disordered regions on D1 and Prod to potentially abrogate the interaction with Piwi will be important for future studies.

      The absence of a TE signal in D1 and Prod mutant ovaries would be much more compelling if investigated more agnostically. The observation that not all TE reporter constructs show a striking signal in the F2 embryos makes me wonder if Burdock and gypsy are not regulated by these two proteins but possibly other TEs are. Alternatively, small RNA-seq would more directly address the question of whether D1 and Prod regulate TEs through the piRNA pathway.

      We completely agree with this comment from the reviewer. We have performed RNA seq on D1 heterozygous (control) and D1 mutant ovaries in a chk26006 background. Since Chk2 arrests germ cell development in response to TE de-repression and DNA damage(Ghabrial and Schüpbach 1999; Moon et al. 2018), we reasoned that the chk2 mutant background would prevent developmental arrest of potential TE-expressing germ cells and we observed that both genotypes exhibited similar gonad morphology (Fig. 4A). From our analysis, we do not observe a significant effect on TE expression in the absence of D1, except for the LTR retrotransposon tirant (Fig. 4B). We also do not observe differential expression of TE repression genes (Fig. 4F).

      We have complemented our RNA seq experiment with small RNA profiling from D1 heterozygous (control) and D1 mutant ovaries. Here, overall piRNA production and antisense piRNAs mapping to TEs were largely unperturbed (Fig. 4C-E).

      Overall, our data is consistent with the TE reporter data (Fig. S7) and suggests that zygotic depletion of D1 does not have a prominent role in TE repression. However, we have uncovered that the presence of D1 during embryogenesis is critical for TE repression in adult gonads (Fig. 6, Fig. S9).

      I had trouble understanding the significance of the Cuff mis-localization when D1 is depleted. Given Cuff's role in the piRNA pathway and close association with chromatin, what would the null hypothesis be for Cuff localization when a chromocenter is disrupted? What is the null expectation of % Cuff at chromocenter given that the chromocenter itself expands massively in size (Figure 4D). The relationship between these two factors seems rather indirect and indeed, the absence of Cuff in the AP would suggest this. The biggest surprise is the absence of TE phenotype in the ovary, given the Cuff mutant phenotype - but we can't rule out given the absence of a genome-wide analysis. I think that these data leave the reader unconvinced that the F2 phenotype is causally linked to Cuff mislocalization.

      We apologize that this data was not more clearly represented. In a wild-type context, Cuff is distributed in a punctate manner across the nurse cell nuclei, with the puncta likely representing piRNA clusters (Fig. 5A-B). We find that a small fraction of Cuff (~5%) is present adjacent to the nurse cell chromocenter (inset, Fig. 5A and Fig. 5D). In the absence of D1, the nurse cell chromocenters increase ~3-4 fold in size. However, the null expectation is still that ~5% of total Cuff would be adjacent to the chromocenter, since the piRNA clusters are not expected to expand in size. In reality, we observe ~27% of total Cuff is mislocalized to chromocenters. Our data indicate that the satellite DNA repeats at the larger chromocenters must be more accessible to Cuff in the D1 mutant nurse cells. This observation is corroborated by the significant increase in piRNAs corresponding to the 1.688 satellite DNA repeat family (Fig. 4E).

      The lack of TE expression in the F1 D1 mutant was indeed surprising based on the Cuff mislocalization but as the reviewers pointed out, we only analyzed two TE reporter constructs in the initial submission. In the revised manuscript, we have performed both RNA seq and small RNA seq. Surprisingly, our data reveal that the Cuff mislocalization does not significantly affect piRNA biogenesis (Fig. 4C, D) and piRNAs mapping to TEs. As a result, both TE repression (Fig. 4B) and fertility (Fig. 6D) are largely preserved in the absence of D1 in adult ovaries.

      Finally, we thank the reviewer for their excellent suggestion to incorporate the F2 D1 heterozygote (Fig. S9) in our analysis! This important control has revealed that the maternal effect of the D1 mutant is negligible for gonad development and fertility (Fig. 6B-D). Rather, our data clearly emphasize embryogenesis (or early larval development) as a key period during which D1 can promote heritable TE repression. Essentially, D1 is not required during adulthood for TE repression if it was present in the early stages of development.

      Apologies if I missed this, but Figure 5 shows the F2 D1 mutant ovaries only. Did you look at the TM6 ovaries as well? These ovaries should lack the maternally provisioned D1 (assuming that females are on the right side) but have the zygotic transcription.

      As mentioned above, this was a great suggestion and we've now characterized this important control in the context of germline development and fertility, to the significant advantage of our paper.

      Minor comments 9. Add line numbers for ease of reference

      We apologize for this. Line numbers have been added in the full revision.

      1. The function of satellite DNA itself is still quite controversial - I would recommend being a bit more careful here - the authors could refer instead to genomic regions enriched for satellite DNA are linked to xyz function (see Abstract line 2 and 7, for example.)

      The abstract has been rewritten and does not directly implicate satellite DNA in a specific cellular function. However, we have taken the reviewer's suggestion in the introduction (line 57-58).

      "Genetic conflicts" in the introduction needs more explanation.

      This sentence has been removed from the introduction in the revised manuscript.

      "In contrast" is not quite the right word. Maybe "However" instead (1st line second paragraph of Intro)

      Done. Line 57 of the revised manuscript.

      Results: what is the motivation for using GSC-enriched testis?

      We use GSC-enriched testes for practical reasons. First, they contain a relatively uniform population of mitotically dividing germ cells unlike regular testes which contain a multitude of post-mitotic germ cells such as spermatocytes, spermatids and sperm. Second, GSC-enriched testes are much larger than normal testes and reduced the number of dissections that were needed for each replicate.

      1. Clarify sentence about the 500 proteins in the Results section - it's not clear from context that this is the union of all experiments.

      Done. Lines 145-149 in the revised manuscript.

      The data reported are not the first to suggest that satellite DNA may have special DNA repair requirements. e.g., PMID: 25340780

      We apologize if we gave the impression that we were making a novel claim. Specialized DNA repair requirements at repetitive sequences have indeed been previously hypothesized(Charlesworth et al. 1994) and studied and we have altered the text to better reflect this (lines 193-195). We have cited the study suggested by the reviewer as well as studies from the Chiolo(Chiolo et al. 2011; Ryu et al. 2015; Caridi et al. 2018) and Soutoglou(Mitrentsi et al. 2022) labs, which have also addressed this fascinating question.

      Page 10: indicate-> indicates.

      Done.

      1. Page 14: revise for clarity: "investigate a context whether these interactions could not take place"

      We've incorporated this suggestion in the revised text (lines 383-386).

      1. Might be important to highlight the 500 interactions are both direct and indirect. "Interacting proteins" alone suggests direct interactions only.

      Done. Lines 145-149.

      The effect of the aub mutant on chromocenter foci did not seem modest to me - however, the bar graphs obscure the raw data - consider plotting all the data not just the mean and error?

      Done. This data is now represented by a box-and-whisker plot (Fig. S5), which shows the distribution of the data.

      Reviewer #2 (Significance (Required)):

      The dataset produced by the AP-Mass Spec offers chromosome biologists an unprecedented resource. The PCH is long-ignored chromosomal region that has historically received minimal attention; consequently, the pathways that regulate heterochromatin are understudied. Moreover, attempting to connect genome organization to transposon regulation is a new and fascinating area. I can easily envision this manuscript triggering a flurry of discovery; however, there is quite a lot of work to do before the data can fully support the claims.

      This manuscript represents a significant contribution to the field of chromosome biology.

      We thank the reviewer for the positive evaluation of our manuscript, and we really appreciate the great suggestion for the F2 D1 heterozygote control! Overall, we believe that our manuscript is substantially improved with the addition of RNA seq, small RNA seq and important genetic controls. Moreover, we are excited by the potential of our study to open new doors in the study of pericentromeric heterochromatin.

      Reviewer #3 (Evidence, reproducibility and clarity (Required)): In the manuscript entitled "Multi-tissue proteomics identifies a link between satellite DNA organization and transgenerational transposon repression", the authors used two satellite DNA-binding proteins, D1 and Prod, as baits to identify chromocenter-associated proteins through quantitative mass spectrometry. The proteomic analysis identified ~500 proteins across embryos, ovaries, and testes, including several piRNA pathways proteins. Depletion of D1 or Prod did not directly contribute to transposon repression in ovary. However, in the absence of maternal and zygotic D1, there was a dramatic increase of agametic ovaries and transgenerational transposon de-repression. Although the study provides a wealth of proteomic data, it lacks mechanistic insights into how satellite DNA organization influence the interactions with other proteins and their functional consequences.

      We thank the reviewer for highlighting that this study will be a valuable resource for future studies on the composition and function of pericentromeric heterochromatin. Based on the reviewer's request for more mechanistic knowledge into how satellite DNA organization affects transposon repression, we have performed RNA seq and small RNA seq, added important genetic controls and significantly improved our text. In the revised manuscript, our data clearly demonstrate that embryogenesis (and potentially early larval development) is a critical time period when D1 contributes to heritable TE repression. Flies lacking D1 during embryogenesis exhibit TE expression in germ cells as adults, which is associated with Chk2-dependent gonadal atrophy. We are particularly excited by these data since TE loci are embedded in the satellite DNA-rich pericentromeric heterochromatin and our study demonstrates an important role for a satellite DNA-binding protein in TE repression.

      Major____ comments 1. While the identification of numerous interactions is significant, it would be helpful to acknowledge that lots of these proteins were known to bind DNA or heterochromatin regions. To strengthen the study, I recommend conducting functional validation of the identified interactions, in addition to the predictions made by Alphfold 2.

      We are happy to take this comment on board. In fact, we believe that the large number of DNA-binding and heterochromatin-associated proteins identified in this study are a sign of quality for the proteomic datasets. In the revised manuscript, we have highlighted known heterochromatin proteins co-purified by D1/Prod in lines 148-151 as well as proteins previously suggested to interact with D1/Prod from high-throughput studies in lines 153-156 (Guruharsha et al. 2011; Tang et al. 2023). In this study, we have focused on the previously unknown associations between D1/Prod and TE repression proteins and functionally validated these interactions as presented in Figures 3-6.

      The observation of transgenerational de-repression is intriguing. However, to better support this finding, it would be better to provide a mechanistic explanation based on the data presented.

      We appreciate this comment from the reviewer, which is similar to major comment #6 raised by reviewer #2. To generate mechanistic insight into the underlying cause of gonad atrophy in the F2 D1 mutant, we have performed RNA seq, small RNA seq and analyzed germline development and fertility in the F2 D1 heterozygous control (Fig. S9).

      For the RNA seq, we used D1 heterozygous (control) and D1 mutant ovaries in a chk26006 background. Since Chk2 arrests germ cell development in response to TE de-repression and DNA damage(Ghabrial and Schüpbach 1999; Moon et al. 2018), we reasoned that the chk2 mutant background would prevent developmental arrest of potential TE-expressing germ cells and we observed that both genotypes exhibited similar gonad morphology (Fig. 4A). From our analysis, we do not observe a significant effect on TE expression in the absence of D1, except for the LTR retrotransposon tirant (Fig. 4B). We also do not observe differential expression of TE repression genes (Fig. 4F).

      We have complemented our RNA seq experiment with small RNA profiling from D1 heterozygous (control) and D1 mutant ovaries. Here, overall piRNA production and antisense piRNAs mapping to TEs were largely unperturbed (Fig. 4C-E). Together, these data are consistent with the TE reporter data (Fig. S7) and suggests that zygotic depletion of D1 does not have a prominent role in TE repression.

      However, we have uncovered that the presence of D1 during embryogenesis is critical for TE repression in adult gonads (Fig. 6, Fig. S9). Essentially, either only maternal deposited D1 (F1 D1 mutant) or only zygotically expressed D1 (F2 D1 het) were sufficient to ensure TE repression and fertility. In contrast, a lack of D1 during embryogenesis (F2 D1 mutant) led to elevated TE expression and Chk2-mediated gonadal atrophy.

      Our results also explain why previous studies have not implicated either D1 or Prod in TE repression, since D1 likely persists during embryogenesis when using depletion approaches such as RNAi-mediated knockdown or F1 generation mutants.

      Minor____ comments 3. Given the maternal effect of the D1 mutant, in Figure 4, I suggest analyzing not only nurse cells but also oocytes to gain a more comprehensive understanding.

      We agree with the reviewer that this experiment can be informative. In the F2 D1 mutant ovaries, germ cell development does not proceed to a stage where oocytes are specified, thus limiting microscopy-based approaches. Nevertheless, we have gauged oocyte quality in all the genotypes using a fertility assay (Fig. 6D) since even ovaries that have a wild-type appearance can produce dysfunctional gametes. In this experiment, we observe that fertility is largely intact in the F1 D1 mutant and F2 D1 heterozygote strains, suggesting that it is the presence of D1 during embryogenesis (or early larval development) that is critical for heritable TE repression and proper oocyte development.

      The conclusion that D1 and Prod do not directly contribute to the repression of transposons needs further analysis from RNA-seq data. Alternatively, the wording could be adjusted to indicate that D1 and Prod are not required for specific transposon silencing, such as Burdock and gypsy.

      Agreed. We have performed RNA-seq in D1 heterozygous (control) and D1 mutant ovaries in a chk26006 background (Fig. 4A, B) as described above.

      As D1 mutation affects Cuff nuclear localization, it would be insightful to sequence the piRNA in the ovaries.

      Agreed. We have performed small RNA profiling from D1 heterozygous (control) and D1 mutant ovaries. Despite the significant mislocalization of the RDC complex, overall piRNA production and antisense piRNAs mapping to TEs were largely unaffected (Fig. 4C-E). However, we observed significant changes in piRNAs mapping to complex satellite DNA repeats (Fig. 4D), but these changes were not associated with a maternal effect on germline development or fertility (F2 D1 heterozygote, Fig. 6B-D).

      **Referee Cross-Commenting**

      I appreciate the valuable insights provided by the other two reviewers regarding this manuscript. I concur with their assessment that substantial improvements are needed before considering this manuscript for publication.

      1. The proteomics data has the potential to be a valuable resource for other scientific community. However, I share the concerns raised by reviewer 1 about the current quality of the data sets. Addressing this, it will augment the manuscript with additional data to show the success of the precipitation. Moreover, as reviewer 2 and I suggested, additional co-IP validations of the IP-MS results are needed to enhance the reliability.

      In the revised manuscript, we have performed multiple experiments to address the quality of the MS datasets based on comments raised by reviewer #1. They are as follows,

      Out of six total AP-MS experiments in this manuscript (D1 x 3, Prod x 2 and Piwi), we observe a strong enrichment of the bait in 5/6 attempts (log2FC between 4-12, Fig. 2A, B, Fig. S2A, Table S1-S3, Table S7). In the D1 testis sample from the initial submission, the lack of a third biological replicate meant that only the p-value (0.07) was not meeting the cutoff. To address this, we have repeated this experiment with an additional biological replicate (Fig. S2A), and our data now clearly show that D1 is also significantly enriched in the testis sample.

      As suggested by the reviewer #1, we have assessed our lysis conditions (450mM NaCl and benzonase) and the solubilization of our baits post-lysis. In Fig. S1D, we have blotted equivalent fractions of the soluble supernatant and insoluble pellet from GFP-Piwi embryos and show that both GFP-Piwi and D1 are largely solubilized following lysis. In Fig. S1E, we also show that our IP protocol works efficiently.

      The only instance in which we do not detect the bait by mass spectrometry is for GFP-Prod pulldown in embryos. Here, one reason could be relatively low expression of GFP-Prod in comparison to GFP-D1 (Fig. S1E), which may lead to technical difficulties in detecting peptides corresponding to Prod. However, we note that Prod IP from embryos co-purified proteins such as Bocks that were previously suggested as Prod interactors from high-throughput studies (Giot et al. 2003; Guruharsha et al. 2011). In addition, Prod IP from embryos also co-purified proteins known to associate with chromocenters such as Hcs(Reyes-Carmona et al. 2011) and Saf-B(Huo et al. 2020). Finally, the concordance between D1 and Prod co-purified proteins from embryo lysates (Fig. 2A, C) suggest that the Prod IP from embryos is of reasonable quality.

      As for the IP-WB validations, we would point out that chromocenters exhibit properties associated with phase separated biomolecular condensates. In our experience, these condensates tend to associate with other proteins/condensates through low affinity or transient interactions that are rarely preserved in IP-WBs, even if there are strong functional relationships. One example is the association between D1 and Prod, which do not pull each other down in an IP-WB (Jagannathan et al. 2019), even though D1 and Prod foci dynamically associate in the nucleus and mutually regulate each other's ability to cluster satellite DNA repeats (Jagannathan et al. 2019). Similarly, IP-WB using GFP-Piwi embryos did not reveal an interaction with D1 (Fig. S4B). However, our extensive functional validations (Figures 4-6) have revealed a critical role for D1 in heritable TE repression.

      I agree with reviewer 2 that the present conclusion is not appropriate regarding D1 and Prod do not contribute to transposon silencing. As reviewer 2 and I suggested, the systematical analysis by using both mRNA-seq and small RNA-seq is required to draw the conclusion.

      Agreed. We have performed RNA seq and small RNA seq as elaborated above. Our conclusions on the role of D1 in TE repression are now much stronger.

      1. The transgenerational phenotype is an intriguing aspect of the study. I agree with reviewer 2 that the inclusion of TM6 ovaries would be a nice control. Further, it is hard to connect this phenotype with the interactions identified in this manuscript. It would be appreciated if the author could provide a mechanistic explanation.

      We have significantly improved these aspects of our study in the revised manuscript. Through the analysis of germline development in the F2 D1 heterozygotes as suggested by reviewer #2, in addition to the recommended RNA seq and small RNA seq, we have now identified embryogenesis (and potentially early larval development) as a time period when D1 makes an important contribution to TE repression. Loss of D1 during embryogenesis leads to TE expression in adult germline cells, which is associated with Chk2-dependent gonadal atrophy. Taken together, we have pinpointed the specific contribution of a satellite DNA-binding protein to transposon repression.

      Reviewer #3 (Significance (Required)):

      Although this study successfully identified several interactions, the authors did not fully elucidate how these interactions contribute to the transgenerational silencing of transposons or ovary development.

      We thank the reviewer for the thoughtful comments and overall positive evaluation of our study, especially the proteomic dataset. We are confident that the revised manuscript has improved our mechanistic understanding of the contribution made by a satellite DNA-binding protein in TE repression.

      References

      Baumgartner L, Handler D, Platzer SW, Yu C, Duchek P, Brennecke J. 2022. The Drosophila ZAD zinc finger protein Kipferl guides Rhino to piRNA clusters eds. D. Bourc'his, K. Struhl, and Z. Zhang. eLife 11: e80067.

      Caridi CP, D'Agostino C, Ryu T, Zapotoczny G, Delabaere L, Li X, Khodaverdian VY, Amaral N, Lin E, Rau AR, et al. 2018. Nuclear F-actin and myosins drive relocalization of heterochromatic breaks. Nature 559: 54-60.

      Charlesworth B, Sniegowski P, Stephan W. 1994. The evolutionary dynamics of repetitive DNA in eukaryotes. Nature 371: 215-220.

      Chiolo I, Minoda A, Colmenares SU, Polyzos A, Costes SV, Karpen GH. 2011. Double-strand breaks in heterochromatin move outside of a dynamic HP1a domain to complete recombinational repair. Cell 144: 732-744.

      Ghabrial A, Schüpbach T. 1999. Activation of a meiotic checkpoint regulates translation of Gurken during Drosophila oogenesis. Nat Cell Biol 1: 354-357.

      Giot L, Bader JS, Brouwer C, Chaudhuri A, Kuang B, Li Y, Hao YL, Ooi CE, Godwin B, Vitols E, et al. 2003. A protein interaction map of Drosophila melanogaster. Science 302: 1727-1736.

      Guruharsha KG, Rual JF, Zhai B, Mintseris J, Vaidya P, Vaidya N, Beekman C, Wong C, Rhee DY, Cenaj O, et al. 2011. A protein complex network of Drosophila melanogaster. Cell 147: 690-703.

      Huo X, Ji L, Zhang Y, Lv P, Cao X, Wang Q, Yan Z, Dong S, Du D, Zhang F, et al. 2020. The Nuclear Matrix Protein SAFB Cooperates with Major Satellite RNAs to Stabilize Heterochromatin Architecture Partially through Phase Separation. Molecular Cell 77: 368-383.e7.

      Jagannathan M, Cummings R, Yamashita YM. 2019. The modular mechanism of chromocenter formation in Drosophila eds. K. VijayRaghavan and S.A. Gerbi. eLife 8: e43938.

      Mitrentsi I, Lou J, Kerjouan A, Verigos J, Reina-San-Martin B, Hinde E, Soutoglou E. 2022. Heterochromatic repeat clustering imposes a physical barrier on homologous recombination to prevent chromosomal translocations. Molecular Cell 82: 2132-2147.e6.

      Moon S, Cassani M, Lin YA, Wang L, Dou K, Zhang ZZ. 2018. A Robust Transposon-Endogenizing Response from Germline Stem Cells. Dev Cell 47: 660-671 e3.

      Pascovici D, Handler DCL, Wu JX, Haynes PA. 2016. Multiple testing corrections in quantitative proteomics: A useful but blunt tool. PROTEOMICS 16: 2448-2453.

      Reyes-Carmona S, Valadéz-Graham V, Aguilar-Fuentes J, Zurita M, León-Del-Río A. 2011. Trafficking and chromatin dynamics of holocarboxylase synthetase during development of Drosophila melanogaster. Molecular Genetics and Metabolism 103: 240-248.

      Ryu T, Spatola B, Delabaere L, Bowlin K, Hopp H, Kunitake R, Karpen GH, Chiolo I. 2015. Heterochromatic breaks move to the nuclear periphery to continue recombinational repair. Nat Cell Biol 17: 1401-1411.

      Tang H-W, Spirohn K, Hu Y, Hao T, Kovács IA, Gao Y, Binari R, Yang-Zhou D, Wan KH, Bader JS, et al. 2023. Next-generation large-scale binary protein interaction network for Drosophila melanogaster. Nat Commun 14: 2162.

    2. Note: This preprint has been reviewed by subject experts for Review Commons. Content has not been altered except for formatting.

      Learn more at Review Commons


      Referee #2

      Evidence, reproducibility and clarity

      Chavan et al. set out to enrich our compendium of pericentric heterochromatin-associated proteins - and to learn some new biology along the way. An ambitious AP-Mass baited with two DNA satellite-binding proteins (D1 and Prod), conducted across embryos, ovaries, and testes, yielded hundreds of candidate proteins putatively engaged at chromocenters. These proteins are enriched for a restricted number of biological pathways, including DNA repair and transposon regulation. To investigate the latter in greater depth, the authors examine D1 and prod mutants for transposon activity changes using reporter constructs for multiple elements. These reporter constructs revealed no transposon activation in the adult ovary, where many proteins identified in the mass spec experiments restrict transposons. However, the authors suggest that the D1 mutant ovaries do show disrupted localization of a key member of a transposon restriction pathway (Cuff), and infer that this mislocalization triggers a striking, transposon derepression phenotype in the F2 ovaries.

      The dataset produced by the AP-Mass Spec offers chromosome biologists an unprecedented resource. The PCH is long-ignored chromosomal region that has historically received minimal attention; consequently, the pathways that regulate heterochromatin are understudied. Moreover, attempting to connect genome organization to transposon regulation is a new and fascinating area. I can easily envision this manuscript triggering a flurry of discovery; however, there is quite a lot of work to do before the data can fully support the claims.

      Major

      1. The introduction requires quite a radical restructure to better highlight the A) importance of the work and B) limit information whose relevance is not clear early in the manuscript. A. Delineating who makes up heterochromatin is a long-standing problem in chromosome biology. This paper has huge potential to contribute to this field; however, it is not the first. Others are working on this problem in other systems, for example PMID:29272703. Moreover, we have some understanding of the distinct pathways that may impact heterochromatin in different tissues (e.g., piRNA biology in ovaries vs the soma). Also, the mutant phenotypes of prod and D1 are different. Fleshing out these three distinct points could help the reader understand what we know and what we don't know about heterochromatin composition and its special biology. Understanding where we are as a field will offer clear predictions about who the interactors might be that we expect to find. For example, given the dramatically different D1 and prod mutant phenotypes (and allele swap phenotypes), how might the interactors with these proteins differ? What do we know about heterochromatin formation differences in different tissues? And how might these differences impact heterochromatin composition? B. The attempt to offer background on the piRNA pathway and hybrid dysgenesis in the Introduction does not work. As a naïve reader, it was not clear why I was reading about these pathways - it is only explicable once the reader gets to the final third of the Results. Moreover, the reader will not retain this information until the TE results are presented many pages later. I strongly urge the authors to shunt the two TE restriction paragraphs to later in the manuscript. They are currently a major impediment to understanding the power of the experiment - which is to identify new proteins, pathways, and ultimately, biology that are currently obscure because we have so little handle on who makes up heterochromatin.
      2. The implications of the failure to rescue female fertility by the tagged versions of both D1 and Prod are not discussed. Consequently, the reader is left uneasy about how to interpret the data.
      3. How were the significance cut-offs determined? Is the p-value reported the adjusted p-value? As a non-expert in AP-MS, I was surprised to find that the p-value, at least according to the Methods, was not adjusted based on the number of tests. This is particularly relevant given the large/unwieldy(?) number of proteins that were identified as signficant in this study. Moreover, the D1 hit in Piwi pull down is actually not significant according to their criteria of p <0.05 (D1 is p=0.05).
      4. How do we know if the lack of overlap across tissues is indeed germline- or soma-specialization rather than noise? To make this inference, conducting some validation would be required. More generally, I was surprised to see no single interactor validated by reciprocal IP-Westerns to validate the Mass-Spec results, though I am admittedly only adjacent to this technique. Note that colocalization, to my mind, does not validate the AP-MS data - in fact, we would a priori predict that piRNA pathway members would co-localize with PCH given the enrichment of piRNA clusters there.
      5. The AlphaFold2 data are very interesting but seem to lack of negative control. Is it possible to incorporate a dataset of proteins that are not predicted to interact physically to elevate the impact of the ones that you have focused on? Moreover, the structural modeling might suggest a competitive interaction between D1 and piRNAs for Piwi. Is this true? And even if not, how does the structural model contribute to your understanding for how D1 engages with the piRNA pathway? The Cuff mislocalization?
      6. The absence of a TE signal in D1 and Prod mutant ovaries would be much more compelling if investigated more agnostically. The observation that not all TE reporter constructs show a striking signal in the F2 embryos makes me wonder if Burdock and gypsy are not regulated by these two proteins but possibly other TEs are. Alternatively, small RNA-seq would more directly address the question of whether D1 and Prod regulate TEs through the piRNA pathway.
      7. I had trouble understanding the significance of the Cuff mis-localization when D1 is depleted. Given Cuff's role in the piRNA pathway and close association with chromatin, what would the null hypothesis be for Cuff localization when a chromocenter is disrupted? What is the null expectation of % Cuff at chromocenter given that the chromocenter itself expands massively in size (Figure 4D). The relationship between these two factors seems rather indirect and indeed, the absence of Cuff in the AP would suggest this. The biggest surprise is the absence of TE phenotype in the ovary, given the Cuff mutant phenotype - but we can't rule out given the absence of a genome-wide analysis. I think that these data leave the reader unconvinced that the F2 phenotype is causally linked to Cuff mislocalization.
      8. Apologies if I missed this, but Figure 5 shows the F2 D1 mutant ovaries only. Did you look at the TM6 ovaries as well? These ovaries should lack the maternally provisioned D1 (assuming that females are on the right side) but have the zygotic transcription.

      Minor

      1. Add line numbers for ease of reference
      2. The function of satellite DNA itself is still quite controversial - I would recommend being a bit more careful here - the authors could refer instead to genomic regions enriched for satellite DNA are linked to xyz function (see Abstract line 2 and 7, for example.)
      3. "Genetic conflicts" in the introduction needs more explanation.
      4. "In contrast" is not quite the right word. Maybe "However" instead (1st line second paragraph of Intro)
      5. Results: what is the motivation for using GSC-enriched testis?
      6. Clarify sentence about the 500 proteins in the Results section - it's not clear from context that this is the union of all experiments.
      7. The data reported are not the first to suggest that satellite DNA may have special DNA repair requirements. e.g., PMID: 25340780
      8. Page 10: indicate-> indicates.
      9. Page 14: revise for clarity: "investigate a context whether these interactions could not take place"
      10. Might be important to highlight the 500 interactions are both direct and indirect. "Interacting proteins" alone suggests direct interactions only.
      11. The effect of the aub mutant on chromocenter foci did not seem modest to me - however, the bar graphs obscure the raw data - consider plotting all the data not just the mean and error?

      Significance

      The dataset produced by the AP-Mass Spec offers chromosome biologists an unprecedented resource. The PCH is long-ignored chromosomal region that has historically received minimal attention; consequently, the pathways that regulate heterochromatin are understudied. Moreover, attempting to connect genome organization to transposon regulation is a new and fascinating area. I can easily envision this manuscript triggering a flurry of discovery; however, there is quite a lot of work to do before the data can fully support the claims.

      This manuscript represents a significant contribution to the field of chromosome biology.

    1. Reviewer #1 (Public Review):

      The question of whether eyespots mimic eyes has certainly been around for a very long time and led to a good deal of debate and contention. This isn't purely an issue of how eyespots work either, but more widely an example of the potential pitfalls of adopting 'just-so-stories' in biology before conducting the appropriate experiments. Recent years have seen a range of studies testing eye mimicry, often purporting to find evidence for or against it, and not always entirely objectively. Thus, the current study is very welcome, rigorously analysing the findings across a suite of papers based on evidence/effect sizes in a meta-analysis.

      The work is very well conducted, robust, objective, and makes a range of valuable contributions and conclusions, with an extensive use of literature for the research. I have no issues with the analysis undertaken. The results and conclusions are compelling. It's probably fair to say that the topic needs more experiments to really reach firm conclusions but the authors do a good job of acknowledging this and highlighting where that future work would be best placed.

    2. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews: 

      Reviewer #1 (Public Review):

      Summary:

      The question of whether eyespots mimic eyes has certainly been around for a very long time and led to a good deal of debate and contention. This isn't purely an issue of how eyespots work either, but more widely an example of the potential pitfalls of adopting 'just-so-stories' in biology before conducting the appropriate experiments. Recent years have seen a range of studies testing eye mimicry, often purporting to find evidence for or against it, and not always entirely objectively. Thus, the current study is very welcome, rigorously analysing the findings across a suite of papers based on evidence/effect sizes in a meta-analysis.

      Strengths:

      The work is very well conducted, robust, objective, and makes a range of valuable contributions and conclusions, with an extensive use of literature for the research. I have no issues with the analysis undertaken, just some minor comments on the manuscript. The results and conclusions are compelling. It's probably fair to say that the topic needs more experiments to really reach firm conclusions but the authors do a good job of acknowledging this and highlighting where that future work would be best placed.

      Weaknesses:

      There are few weaknesses in this work, just some minor amendments to the text for clarity and information.

      We greatly appreciate Reviewer 1’s positive comments on our manuscript. We also revised our manuscript text and a figure in accordance with Reviewer 1’s recommendations.

      Reviewer #2 (Public Review):

      Many prey animals have eyespot-like markings (called eyespots) which have been shown in experiments to hinder predation. However, why eyespots are effective against predation has been debated. The authors attempt to use a meta-analytical approach to address the issue of whether eye-mimicry or conspicuousness makes eyespots effective against predation. They state that their results support the importance of conspicuousness. However, I am not convinced by this.

      There have been many experimental studies that have weighed in on the debate. Experiments have included manipulating target eyespot properties to make them more or less conspicuous, or to make them more or less similar to eyes. Each study has used its own set of protocols. Experiments have been done indoors with a single predator species, and outdoors where, presumably, a large number of predator species predated upon targets. The targets (i.e, prey with eyespot-like markings) have varied from simple triangular paper pieces with circles printed on them to real lepidopteran wings. Some studies have suggested that conspicuousness is important and eye-mimicry is ineffective, while other studies have suggested that more eye-like targets are better protected. Therefore, there is no consensus across experiments on the eye-mimicry versus conspicuousness debate.

      The authors enter the picture with their meta-analysis. The manuscript is well-written and easy to follow. The meta-analysis appears well-carried out, statistically. Their results suggest that conspicuousness is effective, while eye-mimicry is not. I am not convinced that their meta-analysis provides strong enough evidence for this conclusion. The studies that are part of the meta-analysis are varied in terms of protocols, and no single protocol is necessarily better than another. Support for conspicuousness has come primarily from one research group (as acknowledged by the authors), based on a particular set of protocols.

      Furthermore, although conspicuousness is amenable to being quantified, for e.g., using contrast or size of stimuli, assessment of 'similarity to eyes' is inherently subjective. Therefore, manipulation of 'similarity to eyes' in some studies may have been subtle enough that there was no effect.

      There are a few experiments that have indeed supported eye-mimicry. The results from experiments so far suggest that both eye-mimicry and conspicuousness are effective, possibly depending on the predator(s). Importantly, conspicuousness can benefit from eye-mimicry, while eye-mimicry can benefit from conspicuousness.

      Therefore, I argue that generalizing based on a meta-analysis of a small number of studies that conspicuousness is more important than eye-mimicry is not justified. To summarize, I am not convinced that the current study rules out the importance of eye-mimicry in the evolution of eyespots, although I agree with the authors that conspicuousness is important.

      We understand Reviewer 2’s concerns and have addressed them by adding some sentences in the discussion part (L506- 508, L538-L540). In addition, our findings, which were guided by current knowledge, support the conspicuousness hypothesis, but we acknowledge the two hypotheses are not mutually exclusive (L110-112). We also do not reject the eye mimicry hypothesis. As we have demonstrated, there are still several gaps in the current literature and our understanding (L501-553). Our aim is for this research to stimulate further studies on this intriguing topic and to foster more fruitful discussions.

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors):

      Minor comments

      Lines 59/60: "it is possible that eyespots do not involve mimicry of eyes..."

      The sentence was revised (L59). To enhance readability, we have integrated Reviewer 1's suggestions by simplifying the relevant section instead of using the suggested sentence.

      Line 61: not necessarily aposematism. They might work simply through neophobia, unfamiliarity, etc even without unprofitability

      We changed the text in line with the comment from Reviewer 1 (L61-63).

      Lines 62/63 - this is a little hard to follow because I think you really mean both studies of real lepidopterans as well as artificial targets. Need to explain a bit more clearly.

      We provided an additional explanation of our included primary study type (L64-65).

      Lines 93/94 - not quite that they have nothing to do with predator avoidance, but more that any subjective resemblance to eyes is coincidental, or simply as a result of those marking properties being more effective through conspicuousness in their own right.

      Line 94 - similarly, not just aposematism. You explain the possible reasons above on l92 as also being neophobia, etc.

      We agreed with Reviewer 1’s comments and added more explanations about the conspicuousness hypothesis (L96-97). We have also rewritten the sentences that could be misleading to readers (L428).

      Line 96 - this is perhaps a bit misleading as it seems to conflate mechanism and function. The eye mimicry vs conspicuousness debate is largely about how the so-called 'intimidation' function of eyespots works. That is, how eyespots prevent predators from attacking. The deflection hypothesis is a second function of eyespots, which might also work via consciousness or eye mimicry (e.g. if predators try to peck at 'eyes') but has been less central to the mimicry debate.

      The explanations and suggestions from Reviewer 1 are very helpful. We revised this part of our manuscript (L103-108) and Figure 1 and its legend to make it clearer that the eyespot hypothesis and the conspicuousness hypothesis explain anti-predator functions from a different perspective than the deflection hypothesis.

      There is a third function of eyespots too, that being as mate selection traits. Note that Figure 1 should also be altered to reflect these points.

      We wanted to focus on explaining why eyespot patterns can contribute to prey survival. Therefore, we did not state that eyespot patterns function as mate selection traits in this paragraph. Alternatively, we have already mentioned this in the Discussion part (L455-L465) and rewrote it more clearly (L456).

      Were there enough studies on non-avian predators to analyse in any way? 

      We found a few studies on non-avian predators (e.g. fish, invertebrates, or reptiles), but not enough to conduct a meta-analysis.

      Line 171/72 - why? Can you explain, please.

      The reason we excluded studies that used bright or contrasting patterns as control stimuli in our meta-analysis is to ensure comparability across primary studies. We added an explanation in the text (L180-181).

      Line 177 - can you clarify this?

      Without control stimuli, it is challenging to accurately assess the effect of eyespots or other conspicuous patterns on predation avoidance. Control stimuli allow for a comparison of the effect of eyespots or patterns. We added a more detailed explanation to clarify here (L186-188).

      Line 309 - presumably you mean 33 papers, each of which may have multiple experiments? I might have missed it, but how many individual experiments in total? 

      There were 164 individual experiments. We have now added that information in the manuscript (L320).

      Line 320 - paper shaped in a triangle mostly?

      We cannot say that most artificial prey were triangular. After excluding the caterpillar type, 57.4% were triangular, while the remaining 43.6% were rectangular (Figure 2b).

      Line 406: Stevens.

      We fixed this name, thank you (L417).

      Discussion - nice, balanced and thorough. Much of the work done has been in Northern Europe where eyespot species are less common. Perhaps things may differ in areas where eyespots are more prevalent.

      We appreciate Reviewer 1’s kind words and comments. We agree with your comments and reflected them in our manuscript (L542-545).

      Line 477 - True, and predators often have forward-facing eyes making it likely both would often be seen, but a pair of eyes may not be absolutely crucial to avoidance since sometimes a prey animal may only see one eye of a predator (e.g. if the other is occluded, or only one side of the head is visible).

      We were grateful for Reviewer 1's comment. We added a sentence noting that the eyespots do not necessarily have to be in pairs to resemble eyes (L490-L492).

    1. RhmbdH!cq‘sgdqodnokdmnsad‘mfqxvhsgldnqjhbjldntsnesgdf‘ld)hs!rd‘rhdqsnfnaxsgdqtkdrdudmvgdmH!cq‘sgdqmns-?mcrnHtrt‘kkxcn)enkknvhmfsgdo‘sgnekd‘rsqdrhrs‘mbdsg‘s!roqdrdmsdcsnodnokdvgnnbbtoxsgdr‘ldonrhshnmHnbbtoxhmsg‘so‘qshbtk‘qrxrsdl-Sghrhrvgxodnokdlhfgsk‘tfg‘sq‘bhrsnqrdwhrsinjdr dudm vgdm hs l‘jdr sgdl eddk tmbnlenqs‘akd ̃adb‘trd hmsg‘s rhst‘shnm) sn mns k‘tfg ‘mc qhrj adhmf nrsq‘bhydc ax dudqxnmdl‘x l‘jd sgdl eddk lnqd tmbnlenqs‘akd- Sgd d‘rhdrs ̃‘ksgntfgmnsmdbdrr‘qhkxd‘rx ̃bgnhbdhrsnfn‘knmf-Sghrcndrm!sld‘mvdltrs fn‘knmfnqsg‘svdvhkk)nmkxsg‘shevdfn‘knmfvd!kkqtmhmsnkdrrqdrhrs‘mbdsg‘mhevd cnm!s

      This is a big deal with different friend groups, there are always different humors. It's not just about the humor, it's how you want to fit in with them. If you do something that puts you in the spotlight, there would be two ways to deal with it, either the friend group would ignore it or get mad at you. We dislike having the thought that other people dislike us so we try to blend in and do whatever we can to be a part of them even though it would be uncomfortable for us. Humans want to be noticed and have friends, not be left alone.

    1. really learning a lot! -Stacey Vicari

      Just before the final invite, I'd insert a section that speaks to immediate, short-term and long-term wins. You'd structure it in 3 columns titled As soon as you join..., In a month..., in 6 months... That's where you outline what they can achieve/experience/feel at each stage. It's super powerful and hopefully invites readers to opt for the annual membership because you highlight the long-term transformation.

    1. summary

      Speaking of summaries, AI worse than humans at summaries studies show.

      Succinct reason why by David Chisnall:

      LLMs are good at transforms that have the same shape as ones that appear in their training data. They're fairly good, for example, at generating comments from code because code follows common structures and naming conventions that are mirrored in the comments (with totally different shapes of text).

      In contrast, summarisation is tightly coupled to meaning. Summarisation is not just about making text shorter, it's about discarding things that don't contribute to the overall point and combining related things. This is a problem that requires understanding the material, because it's all about making value judgements.

    1. What allows you to stay active and engaged in your work? The simplest way I could answer that would be that I’ve never thought of making art as a career. It’s certainly a job in a sense, but it’s just not a career. It’s a choice you make somewhere down the line about how you’re going to live your life. It doesn’t mean that you don’t have to deal with the same bullshit everybody else has to deal with as far as making a living and all of that, but it’s a shift in your mind where everything you do becomes a part of the same thing. That’s the way I’ve felt about it. Once that decision got made, and I don’t really know when it was, it was probably sometime when I was in school, that’s how I wanted to live my life. That’s pretty much been the throughline from the beginning.
    1. of both race and gender that remained in place—particularly among its women employees known as computers..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }11211Darden’s arrival at Langley coincided with the early days of digital computing. Although Langley could claim one of the most advanced computing systems of the time—an IBM 704, the first computer to support floating-point math—its resources were still limited. For most data analysis tasks, Langley’s Advanced Computing Division relied upon human computers like Darden herself. These computers were all women, trained in math or a related field, and tasked with performing the calculations that determined everything from the best wing shape for an airplane, to the best flight path to the moon. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Aneta SwianiewiczBut despite the crucial roles they played in advancing this and other NASA research, they were treated like unskilled temporary workers.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }11. They were brought into research groups on a project-by-project basis, often without even being told anything about the source of the data they were asked to analyze..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Lena Zlock Most of the engineers, who were predominantly men, never even bothered to learn the computers’ names.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1111.These women computers have only recently.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Michela Banks begun to receive credit for their crucial work, thanks to scholars of the history of computing.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Roujia Wang—and to journalists like Margot Lee Shetterly, whose book, Hidden Figures: The American Dream and the Untold Story of the Black Women Who Helped Win the Space Race,.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Melinda Rossi along with its film adaptation.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Fagana Stone, is responsible for bringing Christine Darden’s story into the public eye.2 Her story, like those of her colleagues, is one of hard work under discriminatory conditions. Each of these women computers was required to advocate for herself—and some, like Darden, chose also to advocate for others. It is because of both her contributions to data science and her advocacy for women that we have chosen to begin our book, Data Feminism, with Darden’s story. For feminism begins with a belief in the “political, social, and economic equality of the sexes,”.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Michela Banks as the Merriam-Webster Dictionary defines the term—as does, for the record, Beyoncé.3 And any definition of feminism also necessarily includes the activist work that is required to turn that belief into reality.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Yolanda Yang. In Data Feminism, we bring these two aspects of feminism together, demonstrating a way of thinking about data, their analysis, and their display, that is informed by this tradition of feminist activism as well as the legacy of feminist critical thought..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1nyah beanAs for Darden, she did not only apply her skills of data analysis to spaceflight trajectories; she also applied them to her own career path..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Yasin Chowdhury After working at Langley for a number of years, she began to notice two distinct patterns in her workplace: men with math credentials were placed in engineering positions, where they could be promoted through the ranks of the civil service, while women with the same degrees were sent to the computing pools, where they languished until they retired or quit.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }211..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Joe Masnyy She did not want to become one of those women, nor did she want others to experience the same fate. So she gathered up her courage and decided to approach the chief of her division to ask him why..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Yasin Chowdhury As Darden, now seventy-five, told Shetterly in an interview for Hidden Figures, his response was sobering: “Well, nobody’s ever complained,” he told Darden. “The women seem to be happy doing that, so that’s just what they do.”.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }21111In today’s world, Darden might have gotten her boss fired—or at least served with an Equal Employment Opportunity Commission complaint. But at the time that Darden posed her question, stereotypical remarks about “what women do” were par for the course..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Roujia Wang In fact, challenging assumptions about what women could or couldn’t do—especially in the workplace—was the central subject of Betty Friedan’s best-selling book, The Feminine Mystique. Published in 1963, The Feminine Mystique.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Jillian McCarten is often credited with starting feminism’s so-called second wave.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Yolanda Yang.4 Fed up with the enforced return to domesticity following the end of World War II, and inspired by the national conversation about equality of opportunity prompted by the civil rights movement, women across the United States began to organize around a wide range of issues, including reproductive rights.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }21 and domestic violence, as well as the workplace inequality and restrictive gender roles that Darden faced at Langley.That said, Darden’s specific experience as a Black woman with a full-time job was quite different than that of a white suburban housewife—the central focus of The Feminine Mystique. And when critics rightly called out Friedan for failing to acknowledge the range of experiences of women in the United States (and abroad), it was women like Darden, among many others, whom they had in mind. In Feminist Theory: From Margin to Center, another landmark feminist book published in 1984, bell hooks puts it plainly: “[Friedan] did not discuss who would be called in to take care of the children and maintain the home if more women like herself were freed from their house labor and given equal access with white men to the professions. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }11She did not speak of the needs of women without men, without children, without homes. She ignored the existence of all non-white women and poor white women..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Melinda Rossi She did not tell readers whether it was more fulfilling to be a maid, a babysitter, a factory worker, a clerk, or a prostitute than to be a leisure-class housewife.”.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Jillian McCarten5In other words, Friedan had failed to consider how those additional dimensions of individual and group identity—like race and class, not to mention sexuality, ability, age, religion, and geography, among many others—intersect with each other to determine one’s experience in the world.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Jayri Ramirez. Although this concept—intersectionality.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }11—did not have a name when hooks described it, the idea that these dimensions cannot be examined in isolation from each other has a much longer intellectual history..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }116 Then, as now, key scholars and activists were deeply attuned to how the racism embedded in US culture.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }2Fagana Stone, Amanda Christopher, coupled with many other forms of oppression, made it impossible to claim a common experience.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Melinda Rossi—or a common movement—for all women everywhere. Instead, what was needed was “the development of integrated analysis and practice based upon the fact that the major systems of oppression are interlocking.”7 These words are from the Combahee River Collective Statement, written in 1978 by the famed Black feminist activist group out of Boston. In this book, we draw heavily from intersectionality and other concepts developed through the work of Black feminist scholars and activists because they offer some of the best ways for negotiating this multidimensional terrain.Indeed, feminism must be intersectional if it seeks to address the challenges of the present moment..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }2Angela Li, Cynthia Lisee We write as two straight, white women based in the United States, with four advanced degrees and five kids between us. We identify as middle-class and cisgender—meaning that our gender identity matches the sex that we were assigned at birth. We have experienced.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Jillian McCarten sexism in various ways at different points of our lives—being women in tech and academia, birthing and breastfeeding babies, and trying to advocate for ourselves and our bodies in a male-dominated health care system. But we haven’t experienced sexism in ways that other women certainly have or that nonbinary people have, for there are many dimensions of our shared identity, as the authors of this book, that align with dominant group positions. This fact makes it impossible for us to speak from experience about some oppressive forces—racism, for example. But it doesn’t make it impossible for us to educate ourselves.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Melinda Rossi and then speak about racism and the role that white people play in upholding it..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Peem Lerdp Or to challenge ableism and the role that abled people play in upholding it. Or to speak about class and wealth inequalities and the role that well-educated, well-off people play in maintaining those..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Fagana Stone Or to believe in the logic of co-liberation. Or to advocate for justice through equity. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1nyah beanIndeed, a central aim of this book is to describe a form of intersectional feminism that takes the inequities of the present moment as its starting point and begins its own work by asking: How can we use data to remake the world?8This is a complex and weighty task, and it will necessarily remain unfinished. But its size and scope need not stop us—or you, the readers of this book—from taking additional steps toward justice. Consider Christine Darden, who, after speaking up to her division chief, heard nothing from him but radio silence. But then, two weeks later, she was indeed promoted.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Amanda Christopher and transferred to a group focused on sonic boom research. In her new position, Darden was able to begin directing her own research projects and collaborate with colleagues of all genders as a peer. Her self-advocacy serves as a model: a sustained attention to how systems of oppression intersect with each other, informed by the knowledge that comes from direct experience. It offers a guide for challenging power and working toward justice.What Is Data Feminism?Christine Darden would go on to conduct groundbreaking research on sonic boom minimization techniques, author more than sixty scientific papers in the field of computational fluid dynamics, and earn her PhD in mechanical engineering—all while “juggling the duties of Girl Scout mom, Sunday school teacher, trips to music lessons, and homemaker,” Shetterly reports. But even as she ascended the professional ranks, she could tell that her scientific accomplishments were still not being recognized as readily as those of her male counterparts; the men, it seemed, received promotions far more quickly.Darden consulted with Langley’s Equal Opportunity Office, where a white woman by the name of Gloria Champine had been compiling a set of statistics about gender and rank. The data confirmed Darden’s direct experience: that women and men.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Jillian McCarten—even those with identical academic credentials, publication records, and performance reviews—were promoted at vastly different rates. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Aneta SwianiewiczChampine recognized that her data could support Darden in her pursuit of a promotion and, furthermore, that these data could help communicate the systemic nature of the problem at hand. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Yuanxi LiChampine visualized the data in the form of a bar chart, and presented the chart to the director of Darden’s division.9 He was “shocked at the disparity,” Shetterly reports, and Darden received the promotion she had long deserved.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }2Angela Li, Fagana Stone.10 Darden would advance to the top rank in the federal civil service, the first Black woman at Langley to do so. By the time that she retired from NASA, in 2007, Darden was a director herself..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Joe Masnyy11Although Darden’s rise into the leadership ranks at NASA was largely the result of her own knowledge, experience, and grit, her story is one that we can only tell as a result of the past several decades of feminist activism and critical thought. It was a national feminist movement that brought women’s issues to the forefront of US cultural politics, and the changes brought about by that movement were vast. They included both the shifting gender roles that pointed Darden in the direction of employment at NASA and the creation of reporting mechanisms.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; } like the one that enabled her to continue her professional rise..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }2Roujia Wang, Seyoon Ahn But Darden’s success in the workplace was also, presumably, the result of many unnamed colleagues and friends who may or may not have considered themselves feminists. These were the people who provided her with community and support.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Melinda Rossi—and likely a not insignificant number of casserole dinners—as she ascended the government ranks. These types of collective efforts have been made increasingly legible, in turn, because of the feminist scholars and activists whose decades of work have enabled us to recognize that labor—emotional as much as physical—as such today.As should already be apparent, feminism has been defined and used in many ways. Here and throughout the book, we employ the term feminism as a shorthand for the diverse and wide-ranging projects that name and challenge sexism and other forces of oppression, as well as those which seek to create more just, equitable, and livable futures. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }312Because of this broadness, some scholars prefer to use the term feminisms, which clearly signals the range of—and, at times, the incompatibilities among—these various strains of feminist activism and political thought. For reasons of readability, we choose to use the term feminism here, but our feminism is intended to be just as expansive. It includes the work of regular folks like Darden and Champine, public intellectuals like Betty Friedan and bell hooks, and organizing groups like the Combahee River Collective, which have taken direct action to achieve the equality of the sexes. It also includes the work of scholars and other cultural critics—like Kimberlé Crenshaw and Margot Lee Shetterly, among many more—who have used writing to explore the social, political, historical, and conceptual reasons behind the inequality of the sexes that we face today.In the process, these writers and activists have given voice to the many ways in which today’s status quo is unjust.12 These injustices are often the result of historical and contemporary differentials of power, including those among men, women, and nonbinary people, as well as those among white women and Black women, academic researchers and Indigenous communities, and people in the Global North and the Global South..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; } Feminists analyze these power differentials so that they can change them..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1athmar al-ghanim Such a broad focus—one that incorporates race, class, ability, and more—would have sounded strange to Friedan or to the white women largely credited for leading the fight for women’s suffrage in the nineteenth century.13 But the reality is that women of color have long insisted that any movement for gender equality must also consider the ways in which privilege and oppression are intersectional..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1nyah beanBecause the concept of intersectionality is essential for this whole book, let’s get a bit more specific. The term was coined by legal theorist Kimberlé Crenshaw in the late 1980s..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1nyah bean14 In law school, Crenshaw had come across the antidiscrimination case of DeGraffenreid v. General Motors. Emma DeGraffenreid was a Black working mother who had sought a job at a General Motors factory in her town. She was not hired and sued GM for discrimination. The factory did have a history of hiring Black people: many Black men worked in industrial and maintenance jobs there. They also had a history of hiring women: many white women worked there as secretaries. These two pieces of evidence provided the rationale for the judge to throw out the case. Because the company did hire Black people and did hire women, it could not be discriminating based on race or gender. But, Crenshaw wanted to know, what about discrimination on the basis of race and gender together? This was something different, it was real, and it needed to be named. Crenshaw not only named the concept, but would go on to explain and elaborate the idea of intersectionality in award-winning books, papers, and talks.15Key to the idea of intersectionality is that it does not only describe the intersecting aspects of any particular person’s identity (or positionalities, as they are sometimes termed).16 It also describes the intersecting forces of privilege and oppression at work in a given society. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }111Oppression involves the systematic mistreatment of certain groups of people by other groups. It happens when power is not distributed equally—when one group controls the institutions of law, education, and culture, and uses its power to systematically exclude other groups while giving its own group unfair advantages (or simply maintaining the status quo).17 In the case of gender oppression, we can point to the sexism, cissexism.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Amanda Christopher, and patriarchy that is evident in everything from political representation to the wage gap to who speaks more often (or more loudly.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Jillian McCarten) in a meeting..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Tegan Lewis18 In the case of racial oppression, this takes the form of racism and white supremacy. Other forms of oppression include ableism, colonialism, and classism. Each has its particular history and manifests differently in different cultures and contexts, but all involve a dominant group that accrues power and privilege at the expense of others. Moreover, these forces of power and privilege on the one hand and oppression on the other mesh together in ways that multiply their effects.The effects of privilege and oppression are not distributed evenly across all individuals and groups, however. For some, they become an obvious and unavoidable part of daily life, particularly for women and people of color and queer people and immigrants: the list goes on. If you are a member of any or all of these (or other) minoritized groups, you experience their effects everywhere, shaping the choices you make (or don’t get to make) each day. These systems of power are as real as rain..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }2Eva Maria Chavez But forces of oppression can be difficult to detect when you benefit from them (we call this a privilege hazard later in the book).d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }2Yolanda Yang, Jillian McCarten. And this is where data come in: it was a set of intersecting systems of power and privilege that Darden was intent on exposing when she posed her initial question to her division chief. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1g mAnd it was that same set of intersecting systems of power and privilege that Darden sought to challenge when she approached Champine. Darden herself didn’t need any more evidence of the problem she faced; she was already living it every day.19 But when her experience was recorded as data and aggregated with others’ experiences, it could be used to challenge institutional systems of power and have far broader impact than on her career trajectory alone..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1111In this way, Darden models what we call data feminism: a way of thinking about data, both their uses and their limits, that is informed by direct experience, by a commitment to action, and by intersectional feminist thought..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Tegan Lewis T.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }11he starting point for data feminism is something that goes mostly unacknowledged in data science: power is not distributed equally in the world. Those who wield power are disproportionately elite, straight, white, able-bodied, cisgender men from the Global North.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Seng Aung Sein Myint.20 The work of data feminism is first to tune into how standard practices in data science serve to reinforce these existing inequalities and second to use data science to challenge and change the distribution of power..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Megan Foesch21 Underlying data feminism is a belief in and commitment to co-liberation: the idea that oppressive systems of power harm all of us, that they undermine the quality and validity of our work, and that they hinder us from creating true and lasting social impact with data science..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1nyah beanWe wrote this book because we are data scientists and data feminists. Although we speak as a “we” in this book, and share certain identities, experiences, and skills, we have distinct life trajectories and motivations for our work on this project. If we were sitting with you right now, we would each introduce ourselves by answering the question: What brings you here today? Placing ourselves in that scenario, here is what we would have to say.Catherine: I am a hacker mama. I spent fifteen years as a freelance software developer and experimental artist, now professor, working on projects ranging from serendipitous news-recommendation systems to countercartography to civic data literacy to making breast pumps not suck. I’m here writing this book because, for one, the hype around big data and AI is deafeningly male and white and technoheroic .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Jillian McCartenand the time is now to reframe that world with a feminist lens. The second reason I’m here is that my recent experience running a large, equity-focused hackathon taught me just how much people like me—basically, well-meaning liberal white people—are part of the problem in struggling for social justice. This book is one attempt to expose such workings of power, which are inside us as much as outside in the world.22Lauren: I often describe myself as a professional nerd. I worked in software development before going to grad school to study English, with a particular focus on early American literature and culture. (Early means very early—like, the eighteenth century.) As a professor at an engineering school, I now work on research projects that translate this history into contemporary contexts. For instance, I’m writing a book about the history of data visualization, employing machine-learning techniques to analyze abolitionist newspapers, and designing a haptic recreation of a hundred-year-old visualization scheme that looks like a quilt. Through projects like these, I show how the rise of the concept of “data” (which, as it turns out, really took off in the eighteenth century.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Jillian McCarten) is closely connected to the rise of our current concepts of gender and race. So one of my reasons for writing this book is to show how the issues of racism and sexism that we see in data science today are by no means new. The other reason is to help translate humanistic thinking into practice and, in so doing, create more opportunities for humanities scholars to engage with activists, organizers, and communities.23We both strongly believe that data can do good in the world. But for it to do so, we must explicitly acknowledge that a key way that power and privilege operate in the world today has to do with the word data itself..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Seng Aung Sein Myint The word dates to the mid-seventeenth century, when it was introduced to supplement existing terms such as evidence and fact..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Tegan Lewis Identifying information as data, rather than as either of those other two terms, served a rhetorical purpose.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Jillian McCarten.24 It converted otherwise debatable information into the solid basis for subsequent claims. But what information needs to become data before it can be trusted? Or, more precisely, whose information needs to become data before it can be considered as fact and acted upon?.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }2Peem Lerdp, Fagana Stone25 Data feminism must answer these questions, too..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }211The story that begins with Christine Darden entering the gates of Langley, passes through her sustained efforts to confront the structural oppression she encountered there, and concludes with her impressive array of life achievements, is a story about the power of data. Throughout her career, in ways large and small, Darden used data to make arguments and transform lives. But that’s not all. Darden’s feel-good biography is just as much a story about the larger systems of power that required data—rather than the belief in her lived experience.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Cynthia Lisee—to perform that transformative work. An institutional mistrust of Darden’s experiential knowledge was almost certainly a factor in Champine’s decision to create her bar chart. Champine likely recognized, as did Darden herself, that she would need the bar chart to be believed..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }11In this way, the alliance between Darden and Champine, and their work together, underscores the flaws and compromises that are inherent in any data-driven project. The process of converting life experience into data always necessarily entails a reduction of that experience.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Tegan Lewis—along with the historical and conceptual burdens of the term. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }11That Darden and Champine were able to view their work as a success despite these inherent constraints underscores even more the importance of listening to and learning from people whose lives and voices are behind the numbers. No dataset or analysis or visualization or model or algorithm is the result of one person working alone. Data feminism can help to remind us that before there are data, there are people—people who offer up their experience to be counted and analyzed, people who perform that counting and analysis, people who visualize the data and promote the findings of any particular project, and people who use the product in the end..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1nyah bean There are also, always, people who go uncounted—for better or for worse.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }11. And there are problems that cannot be represented—or addressed—by data alone. And so data feminism, like justice, must remain both a goal and a process, one that guides our thoughts and our actions as we move forward toward our goal of remaking the world..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }111Data and Power.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Kaiyun ZhengIt took five state-of-the-art IBM System/360 Model 75 machines to guide the Apollo 11 astronauts to the moon. Each was the size of a car and cost $3.5 million dollars. Fast forward to the present. We now have computers in the form of phones that fit in our pockets and—in the case of the 2019 Apple iPhone XR—can perform more than 140 million more instructions per second than a standard IBM System/360..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Kotaro Garvin26 That rate of change is astounding; it represents an exponential growth in computing capacity (figure 0.2a). We’ve witnessed an equally exponential growth in our ability to collect and record information in digital form—and in the ability to have information collected about us (figure 0.2b).Figure 0.2: (a) The time-series chart included in the original paper on Moore’s law, published in 1965, which posited that the number of transistors that could fit on an integrated circuit (and therefore contribute to computing capacity) would double every year. Courtesy of Gordon Moore. (b) Several years ago, researchers concluded that transistors were approaching their smallest size and that Moore’s law would not hold. Nevertheless, today’s computing power is what enabled Dr. Katie Bouman, a postdoctoral fellow at MIT, to contribute to a project that involved processing and compositing approximately five petabytes of data captured by the Event Horizon Telescope to create the first ever image of a black hole. After the publication of this photo in April 2019 showing her excitement—as one of the scientists on the large team that worked for years to capture the image—Bouman was subsequently trolled and harassed online. Courtesy of Tamy Emma Pepin/Twitter.But the act of collecting and recording data about people is not new at all. From the registers of the dead that were published by church officials in the early modern era to the counts of Indigenous populations that appeared in colonial accounts of the Americas, data collection has long been employed as a technique of consolidating knowledge about the people whose data are collected, and therefore consolidating power over their lives..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Sara Blumenstein27 The close relationship between data and power is perhaps most clearly visible in the historical arc that begins with the logs of people captured and placed aboard slave ships, reducing richly lived lives to numbers and names..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }11 It passes through the eugenics movement, in the late nineteenth and early twentieth centuries, which sought to employ data to quantify the superiority of white people over all others. It continues today in the proliferation of biometrics technologies that, as sociologist Simone Browne has shown, are disproportionately deployed to surveil Black bodies..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }28When Edward Snowden, the former US National Security Agency contractor, leaked his cache of classified documents to the press in 2013, he revealed the degree to which the federal government routinely collects data on its citizens—often with minimal regard to legality or ethics..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Natalie Pei Xu29 At the municipal level, too, governments are starting to collect data on everything from traffic movement to facial expressions in the interests of making cities “smarter.”30 This often translates to reinscribing traditional urban patterns of power such as segregation, the overpolicing of communities of color, and the rationing of ever-scarcer city services..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Melinda Rossi31But the government is not alone in these data-collection efforts; corporations do it too—with profit as their guide. The words and phrases we search for on Google, the times of day we are most active on Facebook, and the number of items we add to our Amazon carts are all tracked and stored as data—data that are then converted into corporate financial gain.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }12. The most trivial of everyday actions—searching for a way around traffic, liking a friend’s cat video, or even stepping out of our front doors in the morning—are now hot commodities. This is not because any of these actions are exceptionally interesting (although we do make an exception for Catherine’s cats) but because these tiny actions can be combined with other tiny actions to generate targeted advertisements and personalized recommendations—in other words, to give us more things to click on, like, or buy.32.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Esmeralda OrrinThis is the data economy, and corporations, often aided by academic researchers, are currently scrambling to see what behaviors—both online and off—remain to be turned into data and then monetized. Nothing is outside of datafication.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Melinda Rossi, as this process is sometimes termed—not your search history, or Catherine’s cats, or the butt that Lauren is currently using to sit in her seat. To wit: Shigeomi Koshimizu, a Tokyo-based professor of engineering, has been designing matrices of sensors that collect data at 360 different positions around a rear end while it is comfortably ensconced in a chair.33 He proposes that people have unique butt signatures, as unique as their fingerprints. In the future, he suggests, our cars could be outfitted with butt-scanners instead of keys or car alarms to identify the driver..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Kotaro GarvinAlthough datafication may occasionally verge into the realm of the absurd, it remains a very serious issue. Decisions of civic, economic, and individual importance are already and increasingly being made by automated systems sifting through large amounts of data. For example, PredPol, a so-called predictive policing company founded in 2012 by an anthropology professor at the University of California, Los Angeles, has been employed by the City of Los Angeles for nearly a decade to determine which neighborhoods to patrol more heavily, and which neighborhoods to (mostly) ignore. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Jillian McCartenBut because PredPol is based on historical crime data and US policing practices have always disproportionately surveilled and patrolled neighborhoods of color, the predictions of where crime will happen in the future look a lot like the racist practices of the past..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }3Fagana Stone, Melinda Rossi, Amanda Christopher34 These systems create what mathematician and writer Cathy O’Neil, in Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy, calls a “pernicious feedback loop,” amplifying the effects of racial bias and of the criminalization of poverty that are already endemic to the United States..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Kaiyun ZhengO’Neil’s solution is to open up the computational systems that produce these racist results. Only by knowing what goes in, she argues, can we understand what comes out. This is a key step in the project of mitigating the effects of biased data. Data feminism additionally requires that we trace those biased data back to their source. PredPol and the “three most objective data points” that it employs certainly amplify existing biases, but they are not the root cause.35 The cause, rather, is the long history of the criminalization of Blackness in the United States, which produces biased policing practices, which produce biased historical data, which are then used to develop risk models for the future..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }36 Tracing these links to historical and ongoing forces of oppression can help us answer the ethical question, Should this system exist?37 In the case of PredPol, the answer is a resounding no.Understanding this long and complicated chain reaction is what has motivated Yeshimabeit Milner, along with Boston-based activists, organizers, and mathematicians, to found Data for Black Lives, an organization dedicated to “using data science to create concrete and measurable change in the lives of Black communities.”38 Groups like the Stop LAPD Spying coalition are using explicitly feminist and antiracist methods to quantify and challenge invasive data collection by law enforcement.39 Data journalists are reverse-engineering algorithms and collecting qualitative data at scale about maternal harm.40 Artists are inviting participants to perform ecological maps and using AI for making intergenerational family memoirs.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Melinda Rossi (figure 0.3a).41All these projects are data science. Many people think of data as numbers alone, but data can also consist of words or stories, colors or sounds, or any type of information that is systematically collected, organized, and analyzed .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }12(figures 0.3b, 0.3c).42 The science in data science simply implies a commitment to systematic methods of observation and experiment. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Peem LerdpThroughout this book, we deliberately place diverse data science examples alongside each other. They come from individuals and small groups, and from across academic, artistic, nonprofit, journalistic, community-based, and for-profit organizations. This is due to our belief in a capacious definition of data science, one that seeks to include rather than exclude and does not erect barriers based on formal credentials, professional affiliation, size of data, complexity of technical methods, or other external markers of expertise..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Cynthia Lisee Such markers, after all, have long been used to prevent women from fully engaging in any number of professional fields, even as those fields—which include data science and computer science, among many others—were largely built on the knowledge that women were required to teach themselves.43 An attempt to push back against this gendered history is foundational to data feminism, too.Throughout its own history, feminism has consistently had to work to convince the world that it is relevant to people of all genders.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }2Fagana Stone, Amanda Christopher. We make the same argument: that data feminism is for everybody. (And here we borrow a line from bell hooks.).d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }2Peem Lerdp, Vibha Sathish Kumar44 You will notice that the examples we use are not only about women, nor are they created only by women. That’s because data feminism isn’t only about women. It takes more than one gender to have gender inequality and more than one gender to work toward justice. Likewise, data feminism isn’t only for women. Men, nonbinary, and genderqueer people are proud to call themselves feminists and use feminist thought in their work. Moreover, data feminism isn’t only about gender. Intersectional feminists have keyed us into how race, class, sexuality, ability, age, religion, geography, and more are factors that together influence each person’s experience and opportunities in the world. Finally, data feminism is about power.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Peem Lerdp—about who has it and who doesn’t. Intersectional feminism examines unequal power.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Megan Foesch. And in our contemporary world, data is power too. Because the power of data is wielded unjustly, it must be challenged and changed..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1nyah beanData Feminism in ActionData is a double-edged sword. In a very real sense, data have been used as a weapon by those in power to consolidate their control—over places and things, as well as people. Indeed, a central goal of this book is to show how governments and corporations have long employed data and statistics as management techniques to preserve an unequal status quo. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }3Tegan Lewis, Melinda Rossi, Jillian McCartenWorking with data from a feminist perspective requires knowing and acknowledging this history. To frame the trouble with data in another way: it’s not a coincidence that the institution that employed Christine Darden and enabled her professional rise is the same that wielded the results of her data analysis to assert the technological superiority of the United States over its communist adversaries and to plant an American flag on the moon. But this flawed history does not mean ceding control of the future to the powers of the past. Data are part of the problem, to be sure. But they are also part of the solution. Another central goal of this book is to show how the power of data can be wielded back.Figure 0.3: We define data science expansively in this book—here are three examples. (a) Not the Only One by Stephanie Dinkins (2017), is a sculpture that features a Black family through the use of artificial intelligence. The AI is trained and taught by the underrepresented voices of Black and brown individuals in the tech sector. (b) Researcher Margaret Mitchell and colleagues, in “Seeing through the Human Reporting Bias” (2016), have worked on systems to infer what is not said in human speech for the purposes of image classification. For example, people say “green bananas” but not “yellow bananas” because yellow is implied as the default color of the banana. Similarly, people say “woman doctor” but do not say “man doctor,” so it is the words that are not spoken that encode the bias. (c) A gender analysis of Hollywood film dialogue, “Film Dialogue from 2,000 Screenplays Broken Down by Gender and Age,” by Hanah Anderson and Matt Daniels, created for The Pudding, a data journalism start-up (2017).To guide us in this work, we have developed seven core principles. Individually and together, these principles emerge from the foundation of intersectional feminist thought. Each of the following chapters is structured around a single principle. The seven principles of data feminism are as follows:.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Monserrat PadillaExamine power. Data feminism begins by analyzing how power operates in the world.Challenge power. Data feminism commits to challenging unequal power structures and working toward justice.Elevate emotion and embodiment. Data feminism teaches us to value multiple forms of knowledge, including the knowledge that comes from.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }11 people as living, feeling bodies in the world.Rethink binaries and hierarchies. Data feminism requires us to challenge the gender binary, along with other systems of counting and classification that perpetuate oppression..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Eva Maria ChavezEmbrace pluralism. Data feminism insists that the most complete knowledge comes from synthesizing multiple perspectives, with priority .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }3Eva Maria Chavez, Fagana Stone, Tegan Lewisgiven to local, Indigenous, and experiential ways of knowing.Consider context. Data feminism asserts that data are not neutral or objective. They are the products of unequal social relations, and this context is essential for conducting accurate, ethical analysis..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Natalie Pei XuMake labor visible. The work of data science, like all work in the world, is the work of many hands. .d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Melinda RossiData feminism makes this labor visible so that it can be recognized and valued.Each of the following chapters takes up one of these principles, drawing upon examples from the field of data science, expansively defined, to show how that principle can be put into action. Along the way, we introduce key feminist concepts like the matrix of domination (Patricia Hill Collins; see chapter 1), situated knowledge (Donna Haraway; see chapter 3), and emotional labor (Arlie Hochschild; see chapter 8), as well as some of our own ideas about what data feminism looks like in theory and practice. To this end, we introduce you to people at the cutting edge of data and justice. These include engineers and software developers, activists and community organizers, data journalists, artists, and scholars. This range of people, and the range of projects they have helped to create, is our way of answering the question: What makes a project feminist? As will become clear, a project may be feminist in content, in that it challenges power by choice of subject matter; in form, in that it challenges power by shifting the aesthetic and/or sensory registers of data communication; and/or in process, in that it challenges power by building participatory, inclusive processes of knowledge production.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }11. What unites this broad scope of data-based work is a commitment to action and a desire to remake the world..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Sara BlumensteinOur overarching goal is to take a stand against the status quo—against a world that benefits us, two white college professors, at the expense of others..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Justine Smith To work toward this goal, we have chosen to feature the voices of those who speak from the margins, whether because of their gender, sexuality, race, ability, class, geographic location, or any combination of those (and other) subject positions. We have done so, moreover, because of our belief that those with direct experience of inequality know better than we do about what actions to take next. For this reason, we have attempted to prioritize the work of people in closer proximity to issues of inequality over those who study inequality from a distance..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Natalie Pei Xu In this book, we pay particular attention to inequalities at the intersection of gender and race. This reflects our location in the United States, where the most entrenched issues of inequality have racism at their source. Our values statement, included as an appendix to this book, discusses the rationale for these authorial choices in more detail.Any book involves making choices about whose voices and whose work to include and whose voices and work to omit. We ask that those who find their perspectives insufficiently addressed or their work insufficiently acknowledged view these gaps as additional openings for conversation. Our sincere hope is to contribute in a small way to a much larger conversation, one that began long before we embarked upon this writing process and that will continue long after these pages are through.This book is intended to provide concrete steps to action for data scientists seeking to learn how feminism can help them work toward justice, and for feminists seeking to learn how their own work can carry over to the growing field of data science. It is also addressed to professionals in all fields in which data-driven decisions are being made.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Melinda Rossi, as well as to communities that want to resist or mobilize the data that surrounds them. It is written for everyone who seeks to better understand the charts and statistics that they encounter in their day-to-day lives, and for everyone who seeks to communicate the significance of such charts and statistics to others..d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Peem LerdpOur claim, once again, is that data feminism is for everyone. It’s for people of all genders. It’s by people of all genders. And most importantly: it’s about much more than gender. Data feminism is about power, about who has it and who doesn’t, and about how those differentials of power can be challenged and changed using data.d-undefined, .lh-undefined { background-color: rgba(0, 0, 0, 0.2) !important; }1Yolanda Yang. We invite you, the readers of this book, to join us on this journey toward justice and toward remaking our data-driven world.Connections1 of 2children and siblingsfilterA Translation of this Pubمقدمه: چرا علم داده به فمینیسم احتیاج داردby Catherine D'Ignazio and Lauren KleinShow DescriptionPublished on Mar 07, 2024data-feminism.mitpress.mit.eduDescriptionترجمه توسط امیرحسین پی‌براهA Translation of this PubIntroducción: por qué la ciencia de datos necesita feminismoby Catherine D'Ignazio and Lauren KleinShow DescriptionPublished on Apr 23, 2023data-feminism.mitpress.mit.eduDescriptionDataGénero (Coordinación: Mailén García. Traductoras: Ivana Feldfeber,Sofía García, Gina Ballaben, Giselle Arena y Mariángela Petrizzo. Revisión: Helena Suárez Val.Con la ayuda de Diana Duarte Salinas, Ana Amelia Letelier, y Patricia Maria Garcia Iruegas)Footnotes44LicenseCreative Commons Attribution 4.0 International License (CC-BY 4.0)Comments168 .discussion-list .discussion-thread-component.preview:hover, .discussion-list .discussion-thread-component.expanded-preview { border-left: 3px solid #2D2E2F; padding-left: calc(1em - 2px); } ?Login to discussHappy Polarbear: This passage describing the attitude of most male engineers towards their work is both painfully accurate and poignant, portraying them not as respected individuals deserving recognition for their achievements, but merely as inanimate objects, tools for calculation.?Cynthia Lisee: Such a fertile approach”?Cynthia Lisee: There is somethig immeasurable in lived experience, somethind stat would never reach. data not subject to an ethic of human relations based on "welcoming the Other" are mere abstractions and sources of violence Jamia Williams: Thank you! Reframing is essential when many of these events were deemed “riots” when it was Black folks rising up against various systems.Jamia Williams: Still happening today!?Jillian McCarten: The context in which numbers are collected?Jillian McCarten: The idea that some areas, and therefore some people don’t need to be monitored feels immoral. ?Jillian McCarten: I’ve been thinking about how it’s not what you’re doing but what your goal is, and corporations using our data to make more money off us definitely does not feel the same as collecting data on gender discrimination to stop the practice. ?Jillian McCarten: curious what examples it’s better?Jillian McCarten: It’s interesting what we need evidence to believe, and what we willingly believe without evidence ?Jillian McCarten: the word data origionaly meant to communicate that the fact is confirmed to be true- to shut down disputes ?Jillian McCarten: I love linguistic history, I’d like to learn more about this?Jillian McCarten: Yes, I’m afraid how how biases are baked into AI, and then reinforced ?Jillian McCarten: This reminds me of how priviledge is a lot less visible to those who hold it. ?Jillian McCarten: I wonder if she also had access to data on promotions across race. There’s all kinds of discrimination, and the kinds of data seen as worth collecting also reveal bias. I wonder if the white woman who collected the data focused on gender and missed other identities experiencing discrimination. ?Jillian McCarten: I appreciate how the authors directly state their most salient identities; this should be the norm. Oftentimes when I read a book like this I have to research the authors to learn their identities. Identities always influence the way we think and see the world. ?Jillian McCarten: Compelling quote about power?Jillian McCarten: It’s interesting to me that Darden’s story and the book are the two examples given so far. When I took Into to Women’s Studies in undergrad, this book was heavily criticized for mostly speaking on white feminist issues. I appreciate the author giving a more nuanced intersectional framing in the next paragraph. Jamia Williams: Love to know this! Jamia Williams: And it still far from being accomplished?Jillian McCarten: I’m curious which numbers would help communicate that, and how research can help illustrate the prevelence of this type of sexism. ?Jillian McCarten: This is a compelling example of how in our systems of power some people are seen as more valuable than others, and that likely connects to what data sources are seen as valuable.Jamia Williams: “Hidden figure” Jamia Williams: Thank you! Reframing is essential when many of these events were deemed “riots” when it was Black folks rising up against various systems.Jamia Williams: Still happening today!?Jillian McCarten: I think data is especially important in communicating how segregation persists, and how unofficial segregation is often harder to confront. ?Jillian McCarten: I think it’s important to confront the differences between the image of the US presented and the realities that people live in. I resonate with this statement- growing up I was told over and over how the US is the best place to live, and in the past few years I’ve been learning more about the historical and current harms perpetuated by our government?Jillian McCarten: So many decisions and judgement-calls that go into telling historical events, especially a quick summary like this. I’m glad that this author presents the police this way; I think a lot of authors I’ve read will ignore this reality. ?Amanda Christopher: This is a new term for me! ?Amanda Christopher: This makes me wonder how many women before her advocated for themselves, or if she was the first women at NASA to do so as her supervisor claimed. If she was not, why was her case different? What about the culture of the time at NASA allowed for her to be promoted? If she was the first, what would have happened if other women before her had the courage like Christine to speak up.?Melinda Rossi: Perfect for educators!?Melinda Rossi: I like that the authors are working to offer this knowledge to all.?Melinda Rossi: I like this. Giving credit where credit is due…what a concept!?Melinda Rossi: Ok, here’s the good-for-humanity stuff!?Melinda Rossi: The sad part is that it’s mostly used for financial gains, and not for the good of society/humanity. ?Melinda Rossi: This is sad and terrifying…and yet also seems about right. ?Melinda Rossi: I like this. Data can never capture all and that’s important to remember when we are looking at data and generalizing as if all are spoken for.?Tegan Lewis: This sums up our education system-using data and test scores to maintain the inequity in our school system.?Melinda Rossi: Yes! THIS! + 1 more...?Tegan Lewis: Data is more than numbers. What other data could be gathered in a school system??Tegan Lewis: Does it have to??Tegan Lewis: Would this be considered a misuse of data? Or more of the root of bias??Tegan Lewis: data feminism-can be used to expose inequity and challenge systems of power.Esmeralda Orrin: .Ah, capitalism,’?Tegan Lewis: gender oppression-was evident in the case of Darden?Tegan Lewis: Identity?Tegan Lewis: Would this apply to all forms of sexism, regardless of gender??Amanda Christopher: I would say absolutely, yes. I think one large misconception about feminism is that it only focuses on women, not all genders and sexes.Esmeralda Orrin: somehow I’m not surprised that men know what women are happy doing?Melinda Rossi: Finding a supportive community is key! ?Melinda Rossi: I think this part is so important. Being willing to educate themselves on issues that they might unconsciously contribute to is critical.?Melinda Rossi: We are not a monolith!?Melinda Rossi: bell hooks coming in hot with the truth.?Melinda Rossi: Hidden Figures was (sadly) the first time I had ever heard of Black women at NASA.Fagana Stone: The article could have had more power had the authors also included a note about countless studies that show invaluable contribution of diverse backgrounds and perspectives to innovation and progress. Fagana Stone: Not applicable to all cultures, as there are cultures ruled by matriarchs.?Amanda Christopher: Yes and in those cultures feminism may look differently as feminism is focused on equal rights for all genders. Many of the matriarchical cultures have more than two genders. And just about all societies have some form of gender inequalities.Fagana Stone: Wouldn’t the algorithm update itself as more surveillance data is available rather than fixate on old historical data??Melinda Rossi: That’s a good point. You would think it would be able to update with technology advancing as much as it has. + 1 more...Fagana Stone: In a capitalist country, it should be expected to have wealth inequalities… Not everyone can be wealthy nor can everyone struggle financially. Yes, there are systemic injustices, but it takes all parties involved to improve access to and understand importance of education. Dominated by two political parties running on opposing views, I can’t help but feel very pessimistic about significant progress on these issues in the near future (while the country is enacting backward looking policies and laws). Fagana Stone: “Racism” is a learned concept. Born and raised in Azerbaijan, we did not have a concept of racism, to which I was exposed to after having moved to the states. ?Amanda Christopher: Great point to add to the authors’; that it is “impossible to claim a common experience… for all women, everywhere.”Fagana Stone: It is important to note that men too struggle with sufficient paternity leave. It is critical to shift the thought from women being the only ones fit for childcare role to include men as well.Fagana Stone: Women in some states still fight for their reproductive rights!?Melinda Rossi: Fagana, that’s exactly what I was thinking. Some things change, and some things stay the same. Fagana Stone: Critical lesson in articulating the needs with the hope to identify and operationalize solutions.Fagana Stone: Excellent film! I highly recommend it.Fagana Stone: “The Soviet Union was responsible for launching the first human to space, carrying out the first spacewalk, sending the first woman to space, assembling the first modular space station in orbit around Earth (Mir) — and most of these achievements were accomplished using the same space capsule used in the 1960s.”Fagana Stone: Being from one of the former Soviet Union countries, it is also important to note that the Soviet Union had a more considerable tolerance for risk, hence the advancements mentioned in the field of astronautics. ?Rayon Ston: qKaiyun Zheng: I’ve listened to a podcast before, which is called What happens when an algorithm gets it wrong, In Machines We Trust, MIT Technology Review. It mainly talks about the technology of the use of facial recognition in public and where it can go wrong.The podcast begins with a story about a man who is accused of stealing because a computer matches his photo with a picture of the thief caught on a public camera. But in fact, it was a computer error. The computer can't tell whether the thief is a man or a black man, and the police blindly trust the computer's judgment, and moreover, he says that historically black people steal a lot. And based on the conversation in the podcast, the facial recognition technology isn't perfect, it makes mistakes and matches the wrong people. Such problems are not rare, and involve both privacy violations and potential discrimination.It made me realize that we have a lot more to do in data science.Kaiyun Zheng: We’ve learned about the differences between information and data in the very beginning lessons, and this makes me think about why we emphasize “data” instead of “info” here before the term "feminism".Kaiyun Zheng: The mention of the uneven distribution of power in this book piques my curiosity about how the topic will be addressed. I have previously read a book called "Foundation of Information," which discusses the relationship between power and information. The book suggests that when power is concentrated, the information gathered can sometimes deviate from the truth. As a result, I am curious about how data feminism ensures the authenticity and effectiveness of information collection.Additionally, the information of researching history is also mentioned in the later interview, which makes me curious about how the information of the past can be useful in the present so that it can be used as part of data feminism.Kaiyun Zheng: Intersectionality as a new term which appears after feminism is really interesting. I like how it is introduced here which talks about the example of a black woman since I thought it is the manifestation of a much broader phenomenon in the society. From Google, it is defined as "the interconnected nature of social categorizations such as race, class, and gender, regarded as creating overlapping and interdependent systems of discrimination or disadvantage" which strongly linked to the topic "feminism" (actually closer to equal rights).Each person has multiple identities. For example, I am a university student, an employee at a company, and a kid at home. These are just a few of the many labels that can be applied to an individual, including larger categories such as race, gender, and education. In an information-oriented society, labels can often obscure our understanding of the true nature of things and the individuality of a person can be overlooked. Intersectionality, while still categorizing individuals, does so in a more nuanced manner by connecting multiple labels to form a more specific and accurate representation. This can help individuals overcome challenges and reduce the oppression of vulnerable groups by dominant societal forces.Although from my personal point of view, classifying people is not a very good behavior after all, its emergence also reflects the response to various situations, so as to reduce the oppression of the dominant group of society on the vulnerable group.?Yuanxi Li: It's heartening that the value women create in terms of data has ultimately been validated by data itself, and this result has been achieved through mutual assistance among women.?Yuanxi Li: Intersectionality is an important term that shows how race, class, gender, and other individual characteristics affect with each other?Joe Masnyy: This story has shown the possibilities of this sort of advocation, though as stated early this is clearly not the norm. I appreciate the value of anecdotes such as these, although this text would benefit from hard data to show the scope and magnitude of the issue. Hopefully this is something that is explored further on in the text.?Joe Masnyy: This reality was, in the grand scheme of things, not very long ago. You could argue this still persists even today, with many STEM fields still being largely male in demographics. Even still, women tend to make less than men on average in the exact same fields.?Kotaro Garvin: We have so much more capability then before, but why does it seem like we are not making the same kind of progress? Is it not happening? or is it just unrecognized? ?Kotaro Garvin: I think this is one of the greatest ideas I have ever read, but it also shows why data is so important, everybody is unique but we can still be categorized using data. ?Justine Smith: taking a stand against system that is benefit you?Seng Aung Sein Myint: The decision making process is alway opaque. Hope there is some kind of US federal law which push the school to be a little bit transparent than before. ?Seng Aung Sein Myint: This kind of statistic of average, also make something very simple. No, I am not arguing about this data. ?Seng Aung Sein Myint: Hmm. It is strange to read now. ?Finch Brown: This is such a great line! No wonder someone has already commented on it. I have been thinking a lot recently about how subjective human experiences align and diverge, and how insufficient language and data are in describing experiences. A cool article I just read that reminds me of this is from the New Yorker: How We Should Think About Different Styles of Thinking. One main draw for me in data science is tackling the challenge of most accurately representing data and the stories it tells, given its inescapable constraints.?Yasin Chowdhury: Skill is important everywhere but in a different ways. so its good to have skills. ?Yasin Chowdhury: Without this line the entire story would not exist. But still now a days we do not see that courage specially in black women whoa really talented but chose towards non stem fields because of the difference in ratio. ?Jayri Ramirez: I believe that it is important to understand that it is more than ones gender that can affect the experiences of women. I think this statement is a good description of how there are many dimensions which affect racism and other forms of oppression. ?Roujia Wang: This shows that feminism can meet two kinds of human needs, the first is the detailed technical needs of NASA space agency, and the other is to meet the need of women also need equal status and need the same rights as men to achieve their dreams. In this process, feminism and data science are inextricably linked to each other's achievements.?Seyoon Ahn: As it was discussed in comment above, this part demonstrates the needs of feminism in data science and how not just the individuals but the society as a whole can benefit from data science with an approach of feminism. ?Roujia Wang: In that world, the stereotype of women was that women were not allowed to work in the sciences and that women were more at home with young children and taking care of the family than working outside the home. But such stereotypes prevented many talented women from having a chance to make a career out of it.?Roujia Wang: When people are misogynistic, female scientists contribute to data science research, because women can make up for the shortcomings of men in many ways. Women also use their abilities to change the perception of women in the world?Monserrat Padilla: I am really eager to learn and practice more methodically these principles. The key value in being able to analyze data holistically and seeing the subject matter as a whole at the intersections. Putting these principles into practice will allow for a more complete truth to be available while producing data and/or reading data.?Caroline Hayes: I think it is really moving that they decided to use someone as powerful as Darden’s story to start this textbook. As such a strong, smart women she was able to work in an intellectual field and challenge norms like she did in this instance. In a way she is breaking from the data so commonly released on women in and out of the work field. Instead of becoming one of the computers like 100% of the women before her, she became a part of the 1% who changed it for everyone.?Vibha Sathish Kumar: I agree, this part also resounded with me as well. It also makes you wonder about those other women who were stuck in the same situation for years. Many of those women likely didn’t have access to data or have the means to stand up for themselves in the environment set-up for them. I wonder if this issue is also relevant today, where some women do not have the opportunity to share their experience or have it accounted as data. It takes time to have others recognize their privilege and use it to bring others up - maybe data feminism could be a way to do that. ?Natalie Pei Xu: That is sad to notice that there are still many woman is being ignored and stay silence from some reasons. ?Natalie Pei Xu: First hand resource will be more helpful.?Natalie Pei Xu: This conscious awareness of “product of unequal social relation” is important while collecting, analyzing and concluding, since there is already been a lens filtered the primary source. ?Natalie Pei Xu: Besides using data as a powerful tool to pursuit justice, personal privacy is also a critical concern. ?Natalie Pei Xu: This is very inclusive and thoughtful description about feminism which makes it open up to various people among physical and mental features that aiming at the same thing: justice.Eva Maria Chavez: .Eva Maria Chavez: ecFagana Stone: If we were to focus on collecting unbiased data, then why would the authors even mention “priority” in qualifying it? + 1 more...Eva Maria Chavez: ECEva Maria Chavez: emEva Maria Chavez: collective powerEva Maria Chavez: EMCEva Maria Chavez: ?Kim Martin: test?nyah bean: -?nyah bean: -Fagana Stone: Qualitative data can be so powerful!?nyah bean: -?nyah bean: -?nyah bean: -?nyah bean: -?nyah bean: -?nyah bean: -?nyah bean: yes?nyah bean: -?nyah bean: -?nyah bean: -?nyah bean: -?nyah bean: -?nyah bean: -?nyah bean: -?Yolanda Yang: We should know that “We are under this situation.“?Yolanda Yang: Very personally, I am always shocked by how precise the content they suggest “what I may also interested.“ Also reminds me of Health on the phone, that it reminds us of our next coming period time, and usually also precise.?Melinda Rossi: Yes!?Yolanda Yang: People with privilege cannot recognize, even if they do, they are less likely to make any change, as this would decrease their benefit?Jillian McCarten: One quote that I think of often is “when one has held a position of privilege for so long, equality feels like oppression.” ?Yolanda Yang: “Speak“ and MeToo. Makes it visible.?Yolanda Yang: Looking for equality = we need make efforts ahead to it. Need to uncover it. ?Yolanda Yang: Reminds me of china girl or china head, that used at the beginning of analog films, those are females without names that contribute to film industry, but they were not even supposed to be presented to the audiences.?Yolanda Yang: Even though this has been desegregated for years, it still exists among people’s unconsciousness. ?Jeraldynne Gomez: systematically desgined so that women were stagnant in their positions. The disparity of power and the assertion of such system is correlated as it benefits the men who are implementing it ?Michela Banks: Important Annabel DeLair-Dobrovolny: Converting people into data as a means to assert power and dehumanize the “other”.?Michela Banks: definition ?Michela Banks: At least 50 years later. Why at this time??Michela Banks: power distance between men and women ?Michela Banks: were not recognized for intelligence ?Michela Banks: indicates perception of women in workplace?Michela Banks: note segregation during time of education?Michela Banks: describes environment?ethan chang: Shows how much has changed since then… even though can still be seen to this day.Annabel DeLair-Dobrovolny: Power imbalances contributing to the dehumanization of women in the workplace.?athmar al-ghanim: exactly!!! some individuals have such a negative connotation toward “feminism”. but here, it proves that feminism is just a group of like-minded individuals peacefully going after what they want. all feminists want is change, because for so long, there has been none. and it is about time we stopped neglecting the minority and start appreciating and uplifting them.?athmar al-ghanim: its quite sad to see how barely anything has changed in regard to men having the upper hand in workforces, especially those in STEM related fields. ?athmar al-ghanim: this passage resonates with me as it is a big fear of mine, a woman, going into STEM, that I will constantly have to fight twice as hard as a man, just to show that I am worthy of a position that I am qualified for.?Angela Li: I question how long this took and whether there was an internal fight for Darden to receive her long deserved promotion. The reason being is that I find it hard to believe that the men in power are so readily to accept change in which they lose power or control that benefits them. Earlier in this text, when Darden was working as a calculator with no respect or recognition, her supervisor said that the reason women and men lead such different career paths despite having the same credentials was because no one had ever complained. Through these quotes It sounds like the narrative being pushed is that main reason women are oppressed is because men are unaware of the the disparate treatment and effects of their actions which seems too excusable to not be questioned.Fagana Stone: I read this as the systemic discrimination against women was so normalized that it was essentially on everyone’s blindspot. Having such data showed a trend, a factual analysis that no one could ignore. Also, it takes a lot of courage to challenge the status quo, and these ladies found the way to communicate it to their superiors - through numbers!?Angela Li: I’d like to expand and connect on this idea to reaffirm the highlighted statement. I’m connecting it to to the text “Feminism is for Everybody” by Bell Hooks. In early stages of feminism there were a select few types of feminism that were identified. Of these types there were reformist and visionary feminism. reformist feminism focused mainly on equality with men in the workforce which overshadowed the original radical foundations of contemporary feminism which called for reform and restructuring of society to form a fundamentally anti-sexist nation. while white supremacist capitalist patriarchy suppressed visionary feminism, reformist feminists were also eager to silence them because they could maximize their freedom within the existing system and exploit the lower class of subordinated women.?Cynthia Lisee: Thank you for this important insight?Kat Rohrmeier: The definition of dehumanizing.?Melinda Rossi: Right? Gross.?Aneta Swianiewicz: ?Aneta Swianiewicz: ?Aneta Swianiewicz: ?Aneta Swianiewicz: data to expose inequality?Aneta Swianiewicz: ?g m: “institutional mistrust”?g m: Not only looking @ data, but the how. How was it collected? How has it been processed, and by who??Melinda Rossi: ^^^ Yes! Great point!?g m: Why data is important: challenges privileged hazard by making invisible systems visible.?Lena Zlock: Power dynamics and access to knowledge // needs an equitable foundation, clear statement of relations?Lena Zlock: DH as a countercultural phenomenon?Peem Lerdp: Target goals and audiecnes.?Peem Lerdp: Theme 2?Peem Lerdp: Theme 1?Vibha Sathish Kumar: I find it interesting that the authors mention this explicitly to the readers. A clear stated point that everyone is involved with change. ?Peem Lerdp: Insight on “science” in the phrase data science.?Peem Lerdp: Problems with distinction between what is data and what is information involve deciding who holds the power to make those distinction.Fagana Stone: It is important to add that how we interpret data matters as well.?Peem Lerdp: Def’n?Peem Lerdp: Using data to corroborate lived exp.?Peem Lerdp: Dissociating the identity of the author with the ideas discussed by the author.?Peem Lerdp: Intersectionality and its historic roots.?Peem Lerdp: History of gender inequality in workplace.?Megan Foesch: I think this is such an important lens to have when analyzing the world and what is important. Often times, we get caught up in trivial things that are not important in the bigger picture. We must remind ourselves that issues like justice, race, feminism, equality, and power are all crucial everyday issues that we must solve in order to live as a flourishing community. In order to have justice, each individual must be heard and seen which is currently not happening and needs to. ?Megan Foesch: Throughout this whole article I think that this sentence is one of the most important. The authors reflect on how data feminism is truly about power and how the lack of power between genders signifies that there is an inequality. It is important for us to acknowledge and address this inequality so women can feel as empowered, strong, and safe, as men feel. I think it is also important to point out that data feminism isn’t only for women but “men, nonbinary, and genderqueer people”. In order for a change to be made everyone must accept and acknowledge the imbalance of power that occurs in society. ?Megan Foesch: Before taking this class, I had very rarely heard the term Data Feminism, therefore this idea was somewhat new to me. I am familiar with the ideas of feminism however thinking about feminism from a scientific standpoint is one that can help reinforce popular opinions about lack of equality among genders. It is very difficult to argue something when it is science especially when focusing on systems of power and who holds that power as it is backed by scientific data and evidence.?Nick Klagge: It appears that a word or phrase is missing from the end of this sentence. Perhaps “lived experience” or something like that??Sara Blumenstein: What makes a project feminist??Sara Blumenstein: Data as “consolidating power over lives”?Sara Blumenstein: “Data feminism” as goal and process?Sara Blumenstein: Data vs. fact?Sara Blumenstein: Aggregating data to challenge institutional systems of power?will richardson: This is a very deep statement about feminism. It is also very relevent to the readings.?Sara Blumenstein: Defining “feminism” + 1 more...Data FeminismMIT PressRSSLegalPublished withCommunityData FeminismCollectionDData FeminismPubIntroduction: Why Data Science Needs FeminismcollectionData FeminismCite as D’Ignazio, C., & Klein, L. (2020). Introduction: Why Data Science Needs Feminism. In Data Feminism. Retrieved from https://data-feminism.mitpress.mit.edu/pub/frfa9szdduplicateCopymoreMore Cite OptionsTwitterRedditFacebookLinkedInEmailAuto Generated DownloadPDFWordMarkdownEPUBHTMLOpenDocumentPlain TextJATS XMLLaTeXWhat Is Data Feminism?Data and PowerData Feminism in ActiontickRelease #6Aug 25, 2021 3:54 PMdocument-shareRelease #5Aug 25, 2021 3:22 PMdocument-shareRelease #4Feb 11, 2021 10:25 AMdocument-shareRelease #3Jul 27, 2020 9:43 AMdocument-shareRelease #2Jul 27, 2020 9:42 AMdocument-shareRelease #1Mar 16, 2020 9:12 AMWhat Is Data Feminism?Data and PowerData Feminism in Action(function(){function c(){var b=a.contentDocument||a.contentWindow.document;if(b){var d=b.createElement('script');d.innerHTML="window.__CF$cv$params={r:'8be8b165eed78191',t:'MTcyNTU2NTI0Ni4wMDAwMDA='};var a=document.createElement('script');a.nonce='';a.src='/cdn-cgi/challenge-platform/scripts/jsd/main.js';document.getElementsByTagName('head')[0].appendChild(a);";b.getElementsByTagName('head')[0].appendChild(d)}}if(document.body){var a=document.createElement('iframe');a.height=1;a.width=1;a.style.position='absolute';a.style.top=0;a.style.left=0;a.style.border='none';a.style.visibility='hidden';document.body.appendChild(a);if('loading'!==document.readyState)c();else if(window.addEventListener)document.addEventListener('DOMContentLoaded',c);else{var e=document.onreadystatechange||function(){};document.onreadystatechange=function(b){e(b);'loading'!==document.readyState&&(document.onreadystatechange=e,c())}}}})();error

      This is another example of how we need more women in STEM. There are so many officially desegregated organizations. But segragation is embedded in behavior and that is what needs coaching.

    1. The typical student has at least a twenty-hour-a- weekjob, which is very discouraging—it’s reality, but it’s very discour-aging, because it cuts into my concept of what the requirementsof a college education are; the requirements of a college educa-tion are that you devote all or most of your time to that—that’syour job. And of course you get a group of kids who have noconception of what college is all about . . . who don’t understandlearning or the goal of learning.”

      From what I've understand it's people who want to go to college without needing to work and just focusing on college and enjoying life. They don't need to do part time or so.

    1. Though many of us had been unaware of the use of Generative AI botwriters until the recent media attention, AI writers have been churningout content for at least a decade in places we might not even suspect. Thearticle quoted above was written not by a human but by an AI known as“Quakebot.” Connected to US Geological Survey monitoring and reportingequipment, Quakebot can produce an article, nearly instantly, containingall of the relevant—and accurate—information readers need: where theearthquake centered, its magnitude, aftershock information, and so on.AI writers are far more ubiquitous than most of us recognize. Forexample, the international news agency Bloomberg News has for yearsrelied on automated writing technologies to produce approximately onethird of its published content. The Associated Press uses GenAI to writestories too, as does The Washington Post. Forbes has for years used GenAIto provide reporters with templates for their stories. Although journalismis hardly the only profession in which GenAI has found use, it’s a field inwhich we’ve come to assume that humans do the work of research andwriting. Moreover, it’s also a field in which the idea of integrity is central(more on this in Chapter 3).Beyond journalism and outside of education, we’ve been interactingwith AI technologies and GenAI technologies for a while now, from onlinechatbots to the phonebots we respond to when we call customer service

      Al has been used way more than just for educational purposes. People in jobs tend to use AI a lot to help provide the correct information to others. AI is a technology based tool and all jobs that use technology have a sense of when AI can be helpful in projects or articles and when you should use AI

    1. One reason that culture is difficult to define is that it encompasses all the intangible qualities that make people who they are.

      From my understanding, Culture is more than just meets the eye, and for that, I agree with this statement. Culture can. Be as simple as the way one moves as part of one's lifestyle to as deep as those "intangible qualities" of a person's life. Because it's intangible Some examples of these can be traditions or knowledge about culture.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      This paper conducted a GWAS meta-analysis for COVID-19 hospitalization among admixed American populations. The authors identified four genome-wide significant associations, including two novel loci (BAZ2B and DDIAS), and an additional risk locus near CREBBP using cross-ancestry meta-analysis. They utilized multiple strategies to prioritize risk variants and target genes. Finally, they constructed and assessed a polygenic risk score model with 49 variants associated with critical COVID-19 conditions.

      Strengths:

      Given that most of the previous studies were done in European ancestries, this study provides unique findings about the genetics of COVID-19 in admixed American populations. The GWAS data would be a valuable resource for the community. The authors conducted comprehensive analyses using multiple different strategies, including Bayesian fine mapping, colocalization, TWAS, etc., to prioritize risk variants and target genes. The polygenic risk score (PGS) result demonstrated the ability of the cross-population

      PGS model for COVID-19 risk stratification.

      Thank you very much for the positive comments and the willingness to revise this manuscript.

      Weaknesses:

      (1) One of the major limitations of this study is that the GWAS sample size is relatively small, which limits its power.

      (2) The fine mapping section is unclear and there is a lack of information. The authors assumed one causal signal per locus, and only provided credible sets, but did not provide posterior inclusion probabilities (PIP) for the variants to be causal.

      (3) Colocalization and TWAS used eQTL data from GTEx data, which are mainly from European ancestries. It is unclear how much impact the ancestry mismatch would have on the result. The readers should be cautious when interpreting the results and designing follow-up studies.

      We agree with that the sample size is relatively small. Despite that, it was sufficient to reveal novel risk loci supporting the robustness of the main findings. We have indicated this limitation at the end of the discussion section.

      Thank you for rising this point. As suggested, we have also used SuSIE, which allows to assume more than one causal signal per locus. However, in this case the results were not different from those obtained with the original Bayesian colocalization performed with corrcoverage. Regarding the PIP, at the fine mapping stage we are inclined to put more weight on the functional annotations of the variants in the credible set than on the statistical contributions to the signal. This is the reason why we prefer not to put weight on the PIP of the variants but prioritize variants that were enriched functional annotations.

      This is a good point regarding the lack of diversity in GTEx data. We have also used data from AMR populations (GALA II-SAGE models), although it was only available for blood tissue. Regarding the ancestry mismatch between datasets, several studies have attempted to explore the impact. Gay et al. (PMID: 32912333) studied local ancestry effects on eQTLs from the GTEx consortium and concluded that adjustment of eQTLs by local ancestry only yields modest improvement over using global ancestry (as done in GTEx). Moreover, the colocalization results between adjusting by Local Ancestry and Global Ancestry were not significantly different. Besides, Mogil et al. (PMID: 30096133) observed that genes with higher heritability share genetic architecture between populations. Nevertheless, both studies have evidenced decreased power and poorer predictive performances regarding gene expression because of reduced diversity in eQTL analyses. As consequence of the ancestry mismatch, we now warn the readers that this may compromise signal detection (Discussion, lines 531-533). 

      Reviewer #2 (Public Review):

      This is a genome-wide association study of COVID-19 in individuals of admixed American ancestry (AMR) recruited from Brazil, Colombia, Ecuador, Mexico, Paraguay, and Spain. After quality control and admixture analysis, a total of 3,512 individuals were interrogated for 10,671,028 genetic variants (genotyped + imputed). The genetic association results for these cohorts were meta-analyzed with the results from The Host Genetics Initiative (HGI), involving 3,077 cases and 66,686 controls. The authors found two novel genetic loci associated with COVID-19 at 2q24.2 (rs13003835) and 11q14.1 (rs77599934), and other two independent signals at 3p21.31 (rs35731912) and 6p21.1 (rs2477820) already reported as associated with COVID-19 in previous GWASs. Additional meta-analysis with other HGI studies also suggested risk variants near CREBBP, ZBTB7A, and CASC20 genes.

      Strengths:

      These findings rely on state-of-the-art methods in the field of Statistical Genomics and help to address the issue of a low number of GWASs in non-European populations, ultimately contributing to reducing health inequalities across the globe.

      Thank you very much for the positive comments and the willingness to revise this manuscript.

      Weaknesses:

      There is no replication cohort, as acknowledged by the authors (page 29, line 587), and no experimental validation to assess the biological effect of putative causal variants/genes. Thus, the study provides good evidence of association, rather than causation, between the genetic variants and COVID-19. Lastly, I consider it crucial to report the results for the SCOURGE Latin American GWAS, in addition to its meta-analysis with HGI results, since HGI data has a different phenotype scheme (Hospitalized COVID vs Population) compared to SCOURGE (Hospitalized COVID vs Non-hospitalized COVID).

      We essentially agree with the reviewer in that one of the main limitations of the study is the lack of a replication stage because of the use of all available datasets on a one-stage analysis. To contribute to the interpretation of the findings in the absence of a replication stage, we now assessed the replicability of the novel loci using the Meta-Analysis Model-based Assessment of replicability (MAMBA) approach (PMID: 33785739) and included the posterior probabilities of replication in Table 2. We also explored further the potential replicability of signals in other populations. We agree that the results should be interpreted in terms of associations given the lack of functional validation of main findings, so we have slightly modified the discussion.

      As suggested, the SCOURGE Latin American GWAS summary is now accessible by direct request to the Consortium GitHub repository (https://github.com/CIBERER/Scourge-COVID19) (lines 797-799). We have also included the results from the SCOURGE GWAS analysis for the replication of the 40 lead variants in the Supplementary Table 12. Results from the SCOURGE GWAS for the lead variants in the AMR meta-analysis with HGI were already included in the Supplementary Table 2. As note, we have not been able to conduct the meta-analysis with the same hospitalization scheme as in the HGI study since the population-specific results for those analyses were not publicly released. However, sensitivity analyses included within the supplementary material from the COVID-19 Host Genetics Initiative (2021) stated that there were no significant differences in effects (Odds Ratios) between analyses using population controls or just non-hospitalized COVID-19 patients.

      Reviewer #3 (Public Review):

      Summary:

      In the context of the SCOURGE consortium's research, the authors conduct a GWAS meta-analysis on 4,702 hospitalized individuals of admixed American descent suffering from COVID-19. This study identified four significant genetic associations, including two loci initially discovered in Latin American cohorts. Furthermore, a trans-ethnic meta-analysis highlighted an additional novel risk locus in the CREBBP gene, underscoring the critical role of genetic diversity in understanding the pathogenesis of COVID-19.

      Strengths:

      (1) The study identified two novel severe COVID-19 loci (BAZ2B and DDIAS) by the largest GWAS meta-analysis for COVID-19 hospitalization in admixed Americans.

      (2) With a trans-ethnic meta-analysis, an additional risk locus near CREBBP was identified.

      Thank you very much for the positive comments and the willingness to revise this manuscript.

      Weaknesses:

      (1) The GWAS power is limited due to the relatively small number of cases.

      (2) There is no replication study for the novel severe COVID-19 loci, which may lead to false positive findings.

      We agree with that the sample size is relatively small. Despite that, it was sufficient to reveal novel risk loci supporting the robustness of the main findings. We have indicated this limitation at the end of the discussion section.

      Regarding the lack of a replication study, we now assessed the replicability of the novel loci using the Meta-Analysis Model-based Assessment of replicability (MAMBA) approach (PMID: 33785739). We have included the posterior probabilities of replication in Table 2.

      (3) Significant differences exist in the ages between cases and controls, which could potentially introduce biased confounders. I'm curious about how the authors treated age as a covariate. For instance, did they use ten-year intervals? This needs clarification for reproducibility.

      Thank you for rising this point. Age was included as a continuous variable. This has been now indicated in line 667 (within Material and Methods).

      (4)"Those in the top PGS decile exhibited a 5.90-fold (95% CI=3.29-10.60, p=2.79x10-9) greater risk compared to individuals in the lowest decile". I would recommend comparing with the 40-60% PGS decile rather than the lowest decile, as the lowest PGS decile does not represent 'normal controls'.

      Thank you. In the revised version, the PGS categories was compared following the recommendation (lines 461-463).

      (5) In the field of PGS, it's common to require an independent dataset for training and testing the PGS model. Here, there seems to be an overfitting issue due to using the same subjects for both training and testing the variants.

      We are sorry for the misunderstanding. In fact, we have followed the standard to avoid overfitting of the PGS model and have used different training and testing datasets. The training data (GWAS) was the HGI-B2 ALL meta-analysis, in which our AMR GWAS was not included. The PRS model was then tested in the SCOURGE AMR cohort. However, it is true that we did test the combination of the PRS adding the new discovered variants in the SCOURGE cohort. To avoid potential overfitting by adding the new loci, we have excluded from the manuscript the results on which we included the newly discovered variants.

      (6) The variants selected for the PGS appear arbitrary and may not leverage the GWAS findings without an independent training dataset.

      Again, we are sorry for the misunderstanding. The PGS model was built with 43 variants associated with hospitalization or severity within the HGI v7 results and 7 which were discovered by the GenOMICC consortium in their latest study and were not in the latest HGI release. The variants are included within the Supplementary Table 14, but we have now annotated the discovery GWAS.

      (7) The TWAS models were predominantly trained on European samples, and there is no replication study for the findings as well.

      This is a good point regarding the lack of diversity in GTEx data. We have also used data from AMR populations (GALA II-SAGE models), although it was only available for blood tissue. Regarding the ancestry mismatch between datasets, several studies have attempted to explore the impact. Gay et al. (PMID: 32912333) studied local ancestry effects on eQTLs from the GTEx consortium and concluded that adjustment of eQTLs by local ancestry only yields modest improvement over using global ancestry (as done in GTEx). Moreover, the colocalization results between adjusting by Local Ancestry and Global Ancestry were not significantly different. Besides, Mogil et al. (PMID: 30096133) observed that genes with higher heritability share genetic architecture between populations. Nevertheless, both studies have evidenced decreased power and poorer predictive performances regarding gene expression because of reduced diversity in eQTL analyses. As consequence of the ancestry mismatch, we now warn the readers that this may compromise signal detection (Discussion, lines 531-533). 

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors):

      (1) The authors mentioned the fine mapping method did not converge for the locus in chr 11. I would consider trying a different fine-mapping method (such as SuSiE or FINEMAP). It would be helpful to provide posterior inclusion probabilities (PIP) for the variants in fine mapping results and plot the PIP values in the regional association plots.

      As suggested, we have also used SuSIE, which allows to assume more than one causal signal per locus. However, in this case the results were not different from those obtained with the original Bayesian colocalization performed with corrcoverage. SuSIE’s fine-mapping for chromosome 11 prioritized a single variant, which is likely due to the rare frequency. Thus, we have maintained the fine-mapping as it was originally indicated in the previous version of the manuscript but have now included the credible set in Supplementary Table 6.

      Regarding the PIP, at the fine mapping stage we are inclined to put more weight on the functional annotations of the variants in the credible set than on the statistical contributions to the signal. This is the reason why we prefer not to put weight on the PIP of the variants but prioritize variants that were enriched functional annotations.

      (2) Please provide more detailed information about the VEP and V2G analysis and how to interpret those results. My understanding of V2G is that it includes different sources of information (such as molecular QTLs and chromatin interactions from different tissues/cell types, etc.). It is unclear what sources of information and weight settings were used in the V2G model.

      Thank you for rising this point. As suggested, we have clarified the basis for VEP and V2G and the interpretation (lines 732-743).

      (3) The authors identified multiple genes with different strategies, e.g. FUMA, V2G, COLOC, TWAS, etc. How many genes were found/supported by evidence provided by multiple methods? It could be helpful to have a table summarizing the risk genes found by different strategies, and the evidence supporting the genes. e.g. which genes are found by which methods, and the biological functions of the genes, etc.

      Thank you for rising this point. As suggested, we now added a new figure (Figure 5) to summarize the findings with the multiple methods used.

      (4) It would be helpful to make the code/scripts available for reproducibility.

      As suggested, the SCOURGE Latin American GWAS summary and the analysis scripts (https://github.com/CIBERER/Scourge-COVID19/tree/main/scripts/novel-risk-hosp-AMR-2024) are now accessible in the Consortium GitHub repository (https://github.com/CIBERER/Scourge-COVID19) (lines 806-807).

      (5) The fonts in some of the figures (e.g. Figure 2) are hard to read.

      Thank you. We have now included the figures as SVG files.

      Reviewer #2 (Recommendations For The Authors):

      - The abstract lacks a conclusion sentence.

      Thank you. As suggested, we have included two additional sentences with broad conclusions from the study. We preferred to avoid relying on conclusions related to known or new biological links of the prioritized genes given the lack of functional validation of main findings.

      - Regarding the association analysis (page 27, line 677), I wonder if some of the 10 principal components (PCs) are capturing information about the recruitment areas (countries). It may be relevant to test for multicollinearity among these variables.

      Since we acknowledge that some of the categories might be correlated with a certain PC but not all of them do, we have calculated GVIF values for the main variables to assess the categorical variable as a single entity. The scaled GVIF^1(1/2*Df)) value for the categorical variable is 1.52. Thus, if we square this value, we obtain 2.31, which can be then used for applying usual rule-of-thumb for VIF values.

      - Still on the topic of association analysis, did the authors adjust the logistic model for comorbidities variables from Table 1? Given these comorbidities also have a genetic component and their distribution differs between non-hospitalized vs hospitalized, I am concerned that comorbidities might be confounding the association between genetic variants and COVID.

      We did not adjust by comorbidities since HGI studies were not adjusted either and we aimed to be as aligned as possible with HGI. However, as suggested, we have now tested the association between each of the comorbidities in Table 1 and each of the variants in Table 2, using the comorbidities as dependent variables and adjusting for the main covariables (age, sex, PCs and country of recruitment). None of the variants were significantly associated to the comorbidities (line 333).

      - If I understood correctly, the 49 genetic variants used to develop the polygenic risk score model (PRS) were based on the HGI total sample size (data release 7), which is predominantly of European ancestry. I am concerned about the prediction accuracy in the AMR population (PRS transferability issue).

      We have explored literature in search of other PRS to compare the associated OR in our cohort with ORs calculated in European populations. Horowitz et al. (2022) reported an OR of 1.38 for the top 10% with respect to hospitalization risk in European individuals using a GRS with 12 variants.

      We acknowledge that this might be an issue and is now explained in discussion of the revised version (lines 561-568). However, as this is the first time a PRS for COVID-19 is applied to a relatively large AMR cohort, we believe that this analysis will be of value for further analyses regarding PRS transferability, providing a source for comparison in further studies.    

      - On page 23, line 579, the authors acknowledge their "GWAS is underpowered". This sentence requires a sample/power calculation, otherwise, I suggest using "is likely underpowered".

      Thanks for the input. We have modified the sentence as suggested.

      Reviewer #3 (Recommendations For The Authors):

      I wonder if the authors have an approximate date when the GWAS summary statistic will be available. I reviewed some manuscripts in the past, and the authors claimed they would deposit the data soon, but in fact it would not happen until 2 years later.

      The summary statistics are already available from the SCOURGE Consortium repository https://github.com/CIBERER/Scourge-COVID19 (lines 806-807).

    1. So much seafood was once dismissed as the debris of the sea: eels, snared from the Thames River in 16th-century England and tucked into pies in lieu of meat; clams, eaten by New England colonists only in times of desperation; oysters, offered all-you-can-eat for 6 cents at bars in 19th-century New York City

      It's really crazy how prices changed exponentially like that so the lower class men can't eat such "rich food", But I'm glad they overpriced food that in my opinion just sounds gross except clams, which may taste good with lime

    1. In July, President Grover Cleveland dispatched thousands of American soldiers to break the strike, and a federal court issued a preemptive injunction against Debs and the union’s leadership.

      So President Cleveland sent in the military using tax dollars instead of just passing legislation to meet the union's request? It's clear to see that the government was in the hands of the business owners and not the people who elected them into power.

    1. The present research suggests that even when laptops are used solely to take notes, they may still be impairing learning because their use results in shallower processing. In three studies, we found that students who took notes on laptops performed worse on conceptual questions than students who took notes longhand

      this is important because it's showing the affect of people who take notes on their laptop. i can somewhat say i would agree with this research. coming into college, i thought it was so cool to take notes on my laptop but i did notice that i wasn't retaining any of the information, i was just focused on typing.

    1. at least some of my audience sometimes misunderstand this position um they say well you know to express evil is also part of nature it's also part of the universal mind which is correct um but it is also part of you of the universal mind also part of nature to strive against evil to stop evil and sometimes forcefully if need be because you're not just going to wait for evil to come and barbarize your loved ones and violate truth left and right i think what this understand understanding calls for is not the complete cessation of the use of force when force is the last resort that we have at our at our hands what it calls for is the the end of the notion that the use of force is a form of vengeance

      for - question - nilhism - nondualism - is fighting evil a contradiction? - Rupert Spira - Bernado Kastrup - question - nilhism - how do we prevent falling into?

      question - nondualism - is fighting evil a contradiction? - Pondering this idea raises the question: - Is fighting evil a contradiction? - Do we fall into duality if we fight evil? - Does nonduality imply not creating categories of morality of good and evil? - This question has no answer because - If you understand the question, you are already - a language user - applying some morality - We are already post category and post linguistic - we can never undo this and get back to pre-category and pre-linguistic - Fighting evil cannot conquer it because - in fighting evil, this implies using (deadly) force - deadly force results in death, the most extreme form of suffering - It is tantamount to abuse and justifying death is the greatest act of separation, causing great suffering to the other - In effect, we have the same result as the abuser and this can create a new generation of abused

      question - nilhism - how do we prevent falling into? - Rather, what is needed is to PENETRATE moral relativism / dualism altogether to re-discover the common sacred ground both moral categories are based upon - The use of force as a form of vengeance - is the perpetuation of the abused-abuser cycle

    1. Author response:

      The following is the authors’ response to the original reviews.

      Reviewer #1 (Public Review):

      My main point of concern is the precision of dissection. The authors distinguish cells isolated from the tailbud and different areas in the PSM. They suggest that the cell-autonomous timer is initiated, as cells exit the tailbud.

      This is also relevant for the comparison of single cells isolated from the embryo and cells within the embryo. The dissection will always be less precise and cells within the PSM4 region could contain tailbud cells (as also indicated in Figure 1A), while in the analysis of live imaging data cells can be selected more precisely based on their location. This could therefore contribute to the difference in noise between isolated single cells and cells in the embryo. This could also explain why there are "on average more peaks" in isolated cells (p. 6, l. 7).

      This aspect should be considered in the interpretation of the data and mentioned at least in the discussion. (It does not contradict their finding that more anterior cells oscillate less often and differentiate earlier than more posterior ones.)

      Reviewer #1 rightly points out that selecting cells in a timelapse is more precise than manual dissection. Manual dissection is inherently variable but we believe in general it is not a major source of noise in our experiments. To control for this, we compared the results of 11 manual dissections of the posterior quarter of the PSM (PSM4) with those of the pooled PSM4 data. In general, we did not see large differences in the distributions of peak number or arrest timing that would markedly increase the variability of the pooled data above that of the individual dissections (Figure 1 – supplement figure 7). We have edited the text in the Results to highlight this control experiment (page 6, lines 13-17).

      It is of course possible that we picked up adjacent TB cells when dissecting PSM4, however the reviewer’s assertion that inclusion of TB cells “could also explain why there are "on average more peaks" in isolated cells” is incorrect. Later in the paper we show that cells from the TB have almost identical distributions to PSM4 (mean ± SD, PSM4 4.36 ± 1.44; TB 4.26 ± 1.35; Figure 4 _ supplement 1). Thus, inadvertent inclusion of TB cells while dissecting would in fact not increase the number of peaks.

      Here, the authors focus on the question of how cells differentiate. The reverse question is not addressed at all. How do cells maintain their oscillatory state in the tailbud? One possibility is that cells need external signals to maintain that as indicated in Hubaud et al. 2014. In this regard, the definition of tailbud is also very vague. What is the role of neuromesodermal progenitors? The proposal that the timer is started when cells exit the tailbud is at this point a correlation and there is no functional proof, as long as we do not understand how cells maintain the tailbud state. These are points that should be considered in the discussion.

      The reviewer asks “How do cells maintain their oscillatory state in the tailbud?”. This is a very interesting question, but as recognized by the reviewer, beyond the scope of our current paper.

      We now further emphasize the point “One possibility is that cells need external signals to maintain … as indicated in Hubaud et al. 2014” in the Discussion and added a reference to the review Hubaud and Pourquié 2014 (Signalling dynamics in vertebrate segmentation. Nat Rev Mol Cell Biol 15, 709–721 (2014). https://doi.org/10.1038/nrm3891) (page 18, lines 19-22).

      To clarify the definition of the TB, we have stated more clearly in the results (page 15, lines 8-12) that we defined TB cells as all cells posterior to the notochord (minus skin) and analyzed those that survived

      >5 hours post-dissociation, did not divide, and showed transient Her1-YFP dynamics.

      The reviewer asks: What is the role of neuromesodermal progenitors? In responding to this, we refer to Attardi et al., 2018 (Neuromesodermal progenitors are a conserved source of spinal cord with divergent growth dynamics. Development. 2018 Nov 9;145(21):dev166728. doi: 10.1242/dev.166728).

      Around the stage of dissection in zebrafish in our work, there is a small remaining group of cells characterized as NMPs (Sox2 +, Tbxta+ expression) in the dorsal-posterior wall of the TB. These NMPs rarely divide and are not thought to act as a bipotential pool of progenitors for the elongating axis, as is the case in amniotes, rather contributing to the developing spinal cord. How this particular group of cells behaves in culture is unclear as we did not subdivide the TB tissue before culturing. It would be possible to specifically investigate these NMPs regarding a role in TB oscillations, but given the results of Attardi et al., 2018 (small number of cells, low bipotentiality), we argue that it would not be significant for the conclusions of the current work. To indicate this, we included a sentence and a citation of this paper in the results towards the beginning of the section on the tail bud (page 15, lines 8-12).

      The authors observe that the number of oscillations in single cells ex vivo is more variable than in the embryo. This is presumably due to synchronization between neighbouring cells via Notch signalling in the embryo. Would it be possible to add low doses of Notch inhibitor to interfere with efficient synchronization, while at the same time keeping single cell oscillations high enough to be able to quantify them?

      It is a formal possibility that Delta-Notch signaling may have some impact on the variability in the number of oscillations. However, we argue that the significant amount of cell tracking work required to carry out the suggested experiments would not be justified, considering what has been previously shown in the literature. If Delta-Notch signaling was a major factor controlling the variability of the intrinsic program that we describe, then we would expect that in Delta-Notch mutants the anterior- posterior limits of cyclic gene expression in the PSM would extend beyond those seen in wildtype embryos. Specifically, we might expect to see her1 expression extending more anteriorly in mutants to account for the dramatic increase in the number of cells that have 5, 6, 7 and 8 cycles in culture (Fig. 1E versus Fig. 1I). However, as shown in Holley et al., 2002 (Fig. 5A, B; her1 and the notch pathway function within the oscillator mechanism that regulates zebrafish somitogenesis. Development. 2002 Mar;129(5):1175-83. doi: 10.1242/dev.129.5.1175), the anterior limit of her1 expression in the PSM in DeltaD mutants (aei) is not different to WT. Thus, Delta-Notch signaling may exert a limited control over the number of oscillations, but likely not in excess of one cycle difference.

      In the same direction, it would be interesting to test if variation is decreased, when the number of isolated cells is increased, i.e. if cells are cultured in groups of 2, 3 or 4 cells, for instance.

      This is a great proposal – however the culture setup used here is a wide-field system that doesn’t allow us to accurately follow more than one cell at a time. Cells that adhere to each other tend to crawl over each other, blurring their identity in Z. This is also why we excluded dividing cells in culture from the analysis. Experiments carried out with a customized optical setup will be needed to investigate this in the future.

      It seems that the initiation of Mesp2 expression is rather reproducible and less noisy (+/- 2 oscillation cycles), while the number of oscillations varies considerably (and the number of cells continuing to oscillate after Mesp2 expression is too low to account for that). How can the authors explain this apparent discrepancy?

      The observed tight linkage of the Mesp onset and Her1 arrest argue for a single timing mechanism that is upstream of both gene expression events; indeed, this is one of the key implications of the paper. However, the infrequent dissociation of these events observed in FGF-treated cells suggests that more than one timing pathway could be involved, although there are other interpretations. We’ve added more discussion in the text on one vs multi-timers (page 17, lines 19-23 – page 18, line 1 - 8)., see next point.

      The observation that some cells continue oscillating despite the upregulation of Mesp2 should be discussed further and potential mechanism described, such as incomplete differentiation.

      This is an infrequent (5 out of 54 cells) and interesting feature of PSM4 cells in the presence of FGF. We imagine that this disassociation of clock arrest from mesp on-set timing could be the result of alterations in the thresholds in the sensing mechanisms controlling these two processes. Alternatively - as reviewer 2 argues - it might reflect multiple timing mechanisms at work. We have added a discussion of these alternative interpretations (page 17, lines 19-23 – page 18, line 1 - 8).

      Fig. 3 supplement 3 B missing

      It’s there in the BioRxiv downloadable PDF and full text – but seems to not be included when previewing the PDF. Thanks for the heads up.

      Reviewer #2 (Public Review):

      The authors demonstrate convincingly the potential of single mesodermal cells, removed from zebrafish embryos, to show cell-autonomous oscillatory signaling dynamics and differentiation. Their main conclusion is that a cell-autonomous timer operates in these cells and that additional external signals are integrated to tune cellular dynamics. Combined, this is underlying the precision required for proper embryonic segmentation, in vivo. I think this work stands out for its very thorough, quantitative, single-cell real-time imaging approach, both in vitro and also in vivo. A very significant progress and investment in method development, at the level of the imaging setup and also image analysis, was required to achieve this highly demanding task. This work provides new insight into the biology underlying embryo axis segmentation.

      The work is very well presented and accessible. I think most of the conclusions are well supported. Here a my comments and suggestions:

      The authors state that "We compare their cell-autonomous oscillatory and arrest dynamics to those we observe in the embryo at cellular resolution, finding remarkable agreement."

      I think this statement needs to be better placed in context. In absolute terms, the period of oscillations and the timing of differentiation are actually very different in vitro, compared to in vitro. While oscillations have a period of ~30 minutes in vivo, oscillations take twice as long in vitro. Likewise, while the last oscillation is seen after 143 minutes in vivo, the timing of differentiation is very significantly prolonged, i.e.more than doubled, to 373min in vitro (Supplementary Figure 1-9). I understand what the authors mean with 'remarkable agreement', but this statement is at the risk of being misleading. I think the in vitro to in vivo differences (in absolute time scales) needs to be stated more explicitly. In fact, the drastic change in absolute timescales, while preserving the relative ones, i.e. the number of oscillations a cell is showing before onset of differentiation remains relatively invariant, is a remarkable finding that I think merits more consideration (see below).

      We have changed the text in the abstract (page 1, line 28) to clarify that the agreement is in the relative slowing, intensity increases and peak numbers.

      One timer vs. many timers

      The authors show that the oscillation clock slowing down and the timing of differentiation, i.e. the time it takes to activate the gene mesp, are in principle dissociable processes. In physiological conditions, these are however linked. We are hence dealing with several processes, each controlled in time (and thereby space). Rather than suggesting the presence of ‘a timer’, I think the presence of multiple timing mechanisms would reflect the phenomenology better. I would hence suggest separating the questions more consistently, for instance into the following three:

      a.  what underlies the slowing down of oscillations?

      b.  what controls the timing of onset of differentiation?

      c.  and finally, how are these processes linked?

      Currently, these are discussed somewhat interchangeably, for instance here: “Other models posit that the slowing of Her oscillations arise due to an increase of time-delays in the negative feedback loop of the core clock circuit (Yabe, Uriu, and Takada 2023; Ay et al. 2014), suggesting that factors influencing the duration of pre-mRNA splicing, translation, or nuclear transport may be relevant. Whatever the identity, our results indicate the timer ought to exert control over differentiation independent of the clock.”(page 14). In the first part, the slowing down of oscillations is discussed and then the authors conclude on 'the timer', which however is the one timing differentiation, not the slowing down. I think this could be somewhat misleading.

      To help distinguish the clock’s slowing & arrest from differentiation, we have clarified the text in how we describe our experiments using her1-/-; her7-/- cells (page 10, lines 9-20).

      From this and previous studies, we learn/know that without clock oscillations, the onset of differentiation still occurs. For instance in clock mutant embryos (mouse, zebrafish), mesp onset is still occurring, albeit slightly delayed and not in a periodic but smooth progression. This timing of differentiation can occur without a clock and it is this timer the authors refer to "Whatever the identity, our results indicate the timer ought to exert control over differentiation independent of the clock." (page 14). This 'timer' is related to what has been previously termed 'the wavefront' in the classic Clock and Wavefront model from 1976, i.e. a "timing gradient' and smooth progression of cellular change. The experimental evidence showing it is cell-autonomous by the time it has been laid down,, using single cell measurements, is an important finding, and I would suggest to connect it more clearly to the concept of a wavefront, as per model from 1976.

      We have been explicit about the connection to the clock & wavefront in the discussion (page 17, line 12-17).

      Regarding question a., clearly, the timer for the slowing down of oscillations is operating in single cells, an important finding of this study. It is remarkable to note in this context that while the overall, absolute timescale of slowing down is entirely changed by going from in vivo to in vitro, the relative slowing down of oscillations, per cycle, is very much comparable, both in vivo and in vivo.

      We have now pointed out the relative nature of this phenomenon in the abstract, page 1, line 28.

      To me, while this study does not address the nature of this timer directly, the findings imply that the cell-autonomous timer that controls slowing down is, in fact, linked to the oscillations themselves. We have previously discussed such a timer, i.e. a 'self-referential oscillator' mechanism (in mouse embryos, see Lauschke et al., 2013) and it seems the new exciting findings shown here in zebrafish provide important additional evidence in this direction. I would suggest commenting on this potential conceptual link, especially for those readers interested to see general patterns.

      While we posit that the timer provides positional info to the clock to slow oscillations and instruct its arrest – we do not believe that “the findings imply that the cell-autonomous timer that controls slowing down is, in fact, linked to [i.e., governed by] the oscillations themselves.”. As we show, in her1-/-; her7-/- embryos lacking oscillations, the timing / positional information across the PSM still exists as read-out by Mesp expression. Is this different positional information than that used by the clock? – possibly – but given the tight linkage between Mesp onset and the timing/positioning of clock arrest, both cell-autonomously and in the embryo, we argue that the simplest explanation is that the timing/positional information used by the clock and differentiation are the same. Please see page 10, lines 9-20, as well as the discussion (page 17, lines 19-23; page 18. Lines 1-8 ).

      We agree that the timer must communicate to the clock– but this does not mean it is dependent on the clock for positional information.

      Regarding question c., i.e. how the two timing mechanisms are functionally linked, I think concluding that "Whatever the identity, our results indicate the timer ought to exert control over differentiation independent of the clock." (page 14), might be a bit of an oversimplification. It is correct that the timer of differentiation is operating without a clock, however, physiologically, the link to the clock (and hence the dependence of the timescale of clock slowing down), is also evident. As the author states, without clock input, the precision of when and where differentiation occurs is impacted. I would hence emphasize the need to answer question c., more clearly, not to give the impression that the timing of differentiation does not integrate the clock, which above statement could be interpreted to say.

      As far as we can tell, we do not state that “without clock input, the precision of when and where differentiation occurs is impacted”, and we certainly do not want to give this impression. In contrast, as mentioned above, the her1-/-; her7-/- mutant embryo studies indicate that the lack of a clock signal does not change the differentiation timing, i.e. it does not integrate the clock. Of course, in the formation of a real somite in the embryo, the clock’s input might be expected to cause a given cell to differentiate a little earlier or later so as to be coordinated with its neighbors, for example, along a boundary. But this magnitude of timing difference is within one clock cycle at most, and does not match the large variation seen in the cultured cells that spans over many clock cycles.

      A very interesting finding presented here is that in some rare examples, the arrest of oscillations and onset of differentiation (i.e. mesp) can become dissociated. Again, this shows we deal here with interacting, but independent modules. Just as a comment, there is an interesting medaka mutant, called doppelkorn (Elmasri et al. 2004), which shows a reminiscent phenotype "the Medaka dpk mutant shows an expansion of the her7 expression domain, with apparently normal mesp expression levels in the anterior PSM.". The authors might want to refer to this potential in vivo analogue to their single cell phenotype.

      Thank you, we had forgotten this result. Although we do not agree that this result necessarily means there are two interacting modules, we have included a citation to the paper, along with a discussion of alternative explanations for the dissociation (page 18, lines 2-14).

      One strength of the presented in vitro system is that it enables precise control and experimental perturbations. A very informative set of experiments would be to test the dependence of the cell-autonomous timing mechanisms (plural) seen in isolated cells on ongoing signalling cues, for instance via Fgf and Wnt signaling. The inhibition of these pathways with well-characterised inhibitors, in single cells, would provide important additional insight into the nature of the timing mechanisms, their dependence on signaling and potentially even into how these timers are functionally interdependent.

      We agree and in future experiments we are taking advantage of this in vitro system to directly investigate the effect of signaling cues on the intrinsic timing mechanism.

    1. Almost all good writing begins with terrible first efforts. You need to startsomewhere. Start by getting something -- anything -- down on paper. A friend ofmine says that the first draft is the down draft -- you just get it down. The seconddraft is the up draft -- you fix it up. You try to say what you have to say moreaccurately. And the third draft is the dental draft, where you check every tooth, tosee if it's loose or cramped or decayed, or even, God help us, healthy

      I like this example of how the author described it. I liked this analogy and will try to use it in my writings.

    1. , a robot must protect its own existence

      This is an interesting rule to give to the robots because it makes it so that the robots will fight for their "life" in a similar way that humans do. It seems like this rule was put in place to make sure the robots don't just destroy themselves, but giving them this thought can potentially give the robot a thought that it is valuable and with higher forms of AI, maybe it could develop it's own feelings surrounding their life.

    1. Welcome back.

      And in this lesson, I want to talk about how routing works within a VPC and introduce the internet gateway, which is how we configure a VPC so that data can exit to and enter from the AWS public zone and public internet.

      Now, this lesson will be theory where I'm going to introduce routing and the internet gateway to the architecture behind both those things, as well as jump boxes also known as Bastion hosts.

      In the demo lesson, which immediately follows this one, you'll get the opportunity to implement an internet gateway yourself and the animals for life at VPC and fully configure the VPC with public subnet that allow you to connect to that jump box.

      So let's get started.

      We've got a lot to cover.

      A VPC router is a highly available device which is present in every VPC, both default or custom, which moves traffic from somewhere to somewhere else.

      It runs in all the availability zones that the VPC uses and never needs a way about its availability.

      It simply works.

      The router can be networked in every subnet, and the network is just one address of the subnet.

      By default in a custom VPC, without any other configuration, the VPC router will simply route traffic between subnets in that VPC.

      If an EC2 instance in one subnet wants to communicate with something in another subnet, the VPC router is the thing that moves the traffic between subnets.

      Now, the VPC router is controllable.

      You create route tables which influence what's to do with traffic when it leaves a subnet.

      So just to be clear, the route table that's associated with a subnet defines what the VPC router will do when data leaves that subnet.

      A VPC is created with what's known as a main route table.

      If you don't explicitly associate a custom route table with a subnet, it uses the main route table of the VPC.

      If you do associate a route table that you create with a subnet, then when you associate that, the main route table is disassociated.

      A subnet can only have one route table associated with it at any one time, but a route table can be associated with many subnets.

      A route table looks like this in the user interface.

      In this case, this is the main route table for this specific VPC.

      And a route table is just a list of routes.

      This is one of those routes.

      When traffic leaves the subnet that this route table is associated with, the VPC router reviews the IP packets.

      And remember, I said that a packet had a source address and a destination address, as well as some data.

      The VPC router looks at the destination address of all packets leaving the subnet.

      And when it has that address, it looks at the route table and it identifies all the routes which match that destination address.

      And it does that by checking the destination field at the route.

      This destination field determines what destination the route matches.

      Now, the destination field on a route could match exactly one specific IP address.

      It could be an IP with a /32 prefix.

      And remember, that means that it matches one single IP.

      But the destination field on a route could also be a network match.

      So matching an entire network of which that IP is part.

      Or it could be a default route.

      Remember, for IP version 4, I mentioned that 0.0.0.0.0/0 matches all IP version 4 IP addresses.

      That's known as a default route, a catchall.

      In the case where traffic leaving a subnet only matches one route, then that one route is selected.

      If multiple routes match, so maybe there's a specific /32 IP match, maybe there's a /16 network match, and maybe there's a 0.0.0.0/0 default match, well then the prefix is used as a priority.

      The higher the prefix value, the more specific the route is and the higher priority that that route has.

      So the higher the prefix, all the way up to the highest priority of /32, that is used to select which route applies when traffic leaves a subnet.

      Once we get to the point where a single rule in a route table is selected, either the sole route that applies or the one with the highest priority, then the VPC router forwards that traffic through to its destination, which is determined by the target field on the route.

      And the target field will either point at an AWS gateway, or it will say, as with this example, local.

      And local means that the destination is in the VPC itself, so the VPC router can forward the traffic directly.

      All route tables have at least one route, the local route.

      This matches the VPC side range, and it lets the VPC router know that traffic destined for any IP address in the VPC side range is local and it can be delivered directly.

      If the VPC is also IPv6 enabled, people also have another local route matching the IPv6 side for the VPC.

      As is the case with this example, that bottom route, beginning 2600, that is an IPv6 local route.

      That's the IPv6 side of this specific VPC.

      Now, these local routes can never be updated.

      They're always present, and the local routes always take priority.

      They're the exception to that previous rule about the more specific the route is, the higher the priority.

      Local routes always take priority.

      For the exam, remember the route tables are attached to zero or more subnets.

      A subnet has to have a route table.

      It's either the main route table of the VPC or a custom one that you've created.

      A route table controls what happens to data as it leaves the subnet or subnets that that route table is associated with.

      Local routes are always there, uneditable, and match the VPC IPv4 or VPC side range.

      For anything else, higher, prefix values are more specific than they take priority.

      The way the route works is it matches a destination IP, and for that route, it directs traffic towards a specific target.

      Now, a default route, which I'll talk about shortly, is what happens if nothing else matches.

      Now, an internet gateway is one of the most important add-on features available within a VPC.

      It's a regionally resilient gateway which can be attached to a VPC.

      I've highlighted the words "region" and "resilience" because it always comes up in the exam.

      You do not need a gateway per availability zone.

      The internet gateway's resilient by design.

      One internet gateway will cover all of the availability zones in the region which the VPC is using.

      Now, there's a one-to-one relationship between internet gateways and the VPC.

      A VPC can have no internet gateways which makes it entirely private, or it can have one internet gateway.

      Those are the two choices.

      An internet gateway can be created and not attached to a VPC, so it can have zero attachments, but it can only ever be attached to one VPC at a time, at which point it's valid in all of the availability zones that the VPC uses.

      Now, the internet gateway runs from the border of the VPC and the AWS public zone.

      It's what allows services inside the VPC, which are allocated with public IP version 4 addresses or IP version 6 addresses, to be reached from the internet and to connect to the AWS public zone or the internet.

      Of course, the AWS public zone is used if you're accessing S3, SQS, SNS, or any other AWS public services from the VPC.

      Now, it's a managed gateway and so AWS handles the performance.

      From your perspective as an architect, it simply works.

      Now, using an internet gateway within a VPC, it's not all that complex.

      It's a simplified VPC diagram.

      First, we create and attach an internet gateway to a VPC.

      This means that it's available for use inside the VPC.

      We can use it as a target within route tables.

      So then we create a custom route table and it's within route tables.

      So then we create a custom route table and we associate it with a web subnet.

      Then we add IP version 4 and optionally IP version 6 default routes to the route table with the target being the internet gateway.

      Then finally, we configure the subnet to allocate IP version 4 addresses and optionally IP version 6 by default.

      And at that point, once we've done all of those actions together, the subnet is classified as being a public subnet and any services inside that subnet with public IP addresses can communicate to the internet and vice versa and they can communicate with the AWS public zone as long as there's no other security limitations that are in play.

      Now, don't worry if this seems complex.

      You'll get to experience it shortly in the upcoming demo lesson.

      But before that, I want to talk about how IP version 4 addressing actually works inside the VPC because I've seen quite a few difficult questions on the exam based around IP version 4 addressing and I want to clarify exactly how it works.

      So conceptually, this is how an EC2 instance might look if it's using IP version 4 to communicate with a software update server of some kind.

      So we've got the instance on the left with an internet gateway in between and let's say it's a Linux EC2 instance trying to do some software updates to a Linux update server that's located somewhere in the public internet zone.

      So the instance has a private IP address of let's say 10.16.16.20 and it also has an IP version 4 public address that's assigned to it of 43.250.192.20.

      Only that's not how it really works.

      This is another one of those little details which I try to include in my training courses because it really comes invaluable for the exam.

      What actually happens with public IP version 4 addresses is that they never touch the actual services inside the VPC.

      Instead, when you allocate a public IP version 4 address, for example, to this EC2 instance, a record is created which the internet gateway maintains.

      It links the instance's private IP to its allocated public IP.

      So the instance itself is not configured with that public IP.

      That's why when you make an EC2 instance and allocate it a public IP version 4 address, inside the operating system, it only sees the private IP address.

      Keep this in mind for the exam that there are questions which will try to trip you up on this one.

      For IP version 4, it is not configured in the OS with the public IP address.

      So let's look at the flow of data.

      How does this work?

      Well, when the Linux instance wants to communicate with the Linux software update server, it creates a packet of data.

      Now obviously it probably creates a lot of packets.

      Well, let's focus on one for now because it keeps the diagram nice and simple.

      The packet has a source address of the EC2 instance and a destination address of the Linux software update server.

      So at this point, the packet is not configured with any public addressing.

      This packet would not be routable across the public internet.

      It could not reach the Linux update server.

      That's really important to realize.

      Now the packet leaves the instance and because we've configured a default route, it arrives at the internet gateway.

      The internet gateway sees that this packet is from the EC2 instance because it analyzes the source IP address of that packet.

      And it knows that this instance has an associated public IP version 4 address.

      And so it adjusts the packet.

      It changes the packet's source IP address to the public IP address that's allocated to that instance.

      And this IP address, because it's public, is routable across the internet.

      So the internet gateway then forwards the updated packet onto its destination.

      So as far as the Linux software update server is concerned, it's receiving a packet from a source IP address of 43.250.192.20.

      It knows nothing of the private IP address of the EC2 instance.

      Now on the way back, the inverse happens.

      The Linux software update server wants to send a packet back to our EC2 instance.

      But as far as it's concerned, it doesn't know about the private address.

      It just knows about the public address.

      So the software update server sends a packet back, addressed to the instance's public IP address with its source address.

      So it thinks that the real IP address of the instance is this 43.250.192.20 address.

      Now this IP address actually belongs to the internet gateway.

      And so it travels over the public internet and it arrives at the internet gateway, which then modifies this packet.

      It changes it.

      It changes the destination address of this packet from the 43 address to the original IP address of the EC2 instance.

      And it does this because it's got a record of the relationship between the private IP and the allocated public IP.

      So it just changes the destination to the private IP address of the instance.

      And then it forwards this packet through the VPC network to the original EC2 instance.

      So the reason I wanted to highlight this is because at no point is the operating system on the EC2 instance aware of its public IP.

      It just has a private IP.

      Don't fall for any exam questions which try to convince you to assign the public IP version 4 address of an EC2 instance directly to the operating system.

      It has no knowledge of this public address.

      Configuring an EC2 instance appropriately using IP version 4 means putting the private IP address only.

      The public address never touches the instance.

      For IP version 6, all addresses that AWS users are natively, publicly, routable.

      And so in the case of IP version 6, the operating system does have the IP address version 6 address configured upon it.

      That's the publicly routable address.

      And all the internet gateway does is pass traffic from an instance to an internet server.

      And then back again, it doesn't do any translation.

      Now, before we implement this in the demo lesson, I just want to briefly touch upon bastion hosts and jump boxes.

      At a high level, bastion hosts and jump boxes are one and the same.

      Essentially, it's just an instance in a public subnet inside of VPC.

      And architecturally, they're used to allow incoming management connections.

      So all incoming management connections arrive at bastion hosts or jump box.

      And then once connected, you can then go on to access internal only VPC resources.

      So bastion hosts and jump boxes are generally used either as a management point or as an entry point for private only VPCs.

      So if your VPC is a highly secure private VPC, you'll generally have a bastion host or a jump box being the only way to get access to that VPC.

      So it's essentially just an inbound management point.

      And you can configure these bastion hosts or jump boxes to only accept connections from certain IP addresses, to authenticate with SSH, or to integrate with your on-premises identity servers.

      You can configure them exactly how you need, but at a real core architectural level, they are generally the only entry point to a highly secure VPC.

      And historically, they were the only way to manage private VPC instances.

      Now, there are alternative ways to do that now, but you will still find bastion hosts and jump boxes do feature on the exam.

      OK, so that's all the theory that I wanted to cover in this lesson.

      It's now time for a demo.

      In the next lesson, we're going to implement the Internet gateway in the Animals for Life VPC.

      We'll create it, we'll attach it to the VPC, we'll create a custom route table for the web subnets, we'll create two routes in that route table, one for IP version 4 and one for IP version 6.

      And both of these will point at the Internet gateway as a target.

      We'll associate that route table with the web tier subnets, configure those subnets to allocate public IP version 4 addresses, and then launch a bastion host into one of those web subnets.

      And if all goes well, we will be able to connect to that instance using our SSH application.

      So I think this demo lesson is going to be really interesting and really exciting.

      It's the first time that we're going to be stepping through something together that we could qualify as production like.

      Something that you could implement and would implement in a production ready VPC.

      So go ahead, complete this video, and when you're ready, join me in the demo lesson.

    1. Welcome back and in this demo lesson you're going to create all of the subnets within the custom VPC for animals for life.

      So we're going to create the subnets as shown on screen now.

      We're going to be creating four subnets in each availability zone.

      So that's the web subnet, the application subnet, the database subnet and the reserved subnet.

      And we're going to create each of those four in availability zone A, B and C.

      Now before we get started, attached to this lesson is a link to this document.

      Now this is a list of all the details of the subnet you're going to create in this demo lesson.

      So we've got a reserved subnet, a database subnet, an app subnet and a web subnet.

      And we've got each of those in availability zone A, B and C.

      Now in terms of what this document contains, we have a subnet name, then we have the IP range that subnet will use, the availability zone that subnet is within and then this last value and we'll talk about what this is very shortly.

      This relates to IP version 6.

      Now you'll notice that for each subnet this is a unique value.

      You're going to need this to configure the IP version 6 configuration for each of the subnets.

      Now let's get started.

      So let's move to the AWS console and you need to be within the VPC console.

      So if you're not already there, go ahead and type that in the search box at the top and then click to move to the VPC console and then click on subnets.

      And once you've done that, go ahead and click on create subnet.

      Now the newest version of the user interface allows you to create multiple subnets at a time and so we're going to create all four of the subnets in each availability zone.

      So we'll do this three times, one for availability zone A, one for B and one for C.

      So we'll get started with availability zone A and first we need to select the VPC to use.

      So click in the VPC ID drop down and select the animals for live VPC.

      Once you've done that, we're going to start with subnet one of one.

      So let's move to the subnet's document.

      So it's these subnets that we're going to create and we'll start with the reserved subnet.

      So copy the name of the subnet into your clipboard and paste it into subnet name.

      Then change the availability zone to AZA and then make sure IPv4 is set to manual and move back to the subnet's document and copy the IP range that we're going to use and paste that into this box.

      Then scroll down again and make sure manual input is selected for IPv6.

      Then click the drop down and select the IPv6 range for the VPC.

      Now the VPC uses a /56 IPv6 range.

      Because we need our subnet to fit inside this, we're going to make the individual subnet ranges much smaller.

      So what I'll need you to do and you'll need to do this each time is to click on the down arrow and you'll need to click on the down arrow twice.

      The first time will change it to a /60 and the second time to a /64.

      Now note in my case how I have 9, 6, 0, 0 and if I click on this right arrow, it increments this value by one each time.

      Now this value corresponds to the value in the subnet's document, so in this case 0, 0.

      By changing this value each time, it means that you're giving a unique IPv6 range to each subnet.

      So in this case, start off by leaving this set to 0, 0.

      And once you've done that, you can click on add new subnet.

      And we're going to create the next subnet.

      So in this case, it's sn-db-a, so enter that name, change the availability zone to A, manual input for IPv4, copy the IP range for DBA into your clipboard, paste that in, manual for IPv6, change the VPC range in the drop down, click the down arrow twice to set this to /64, and then change this value to 0, 1.

      And again, this matches the IPv6 value in the subnet's document.

      Then we'll do the same process for the third subnet, so we're going to add a new subnet.

      This time the name is sn-app-a, enter that, availability zone A, manual for IPv4, and then paste in the IP range, manual for IPv6, and select the VPC range, and then change the subnet block to /64 by clicking the down arrow twice.

      And then click the right arrow twice to set 0, 2 as the unique value for the subnet.

      And again, this matches the value in the subnet's document.

      Then lastly, we're going to do the same thing for the last subnet in availability zone A, so we're going to add a new subnet.

      This time it's web A, so copy and paste that into the subnet name box, set availability zone A, manual for IPv4, copy and paste the range from the subnet's document, manual for IPv6, select the IPv6 range from the VPC, click the down arrow twice to set the appropriate size for the subnet IPv6 range, and then click on the right arrow to change this value to 0, 3.

      Now that's all the subnet's created, all four of them in availability zone A, so we can scroll all the way down to the bottom, and click on create subnet, and that's going to create all four of those subnets.

      We can see those in this list.

      Now we're going to follow that same process for availability zone B, so click on create subnet, change the VPC to the animals for IPvc, and now we're going to start moving through quicker.

      So first we're going to do the reserve B subnet, so copy the name, paste that in, set the availability zone this time to B, manual for IPv4, paste in the range, manual for IPv6, select the VPC range in the drop-down, click on the down arrow twice to set the /64, and then click on the right arrow and set to 0, 4 as the unique value.

      Scroll down, click add new subnet, next is DBB, enter that, availability zone B, manual for IPv4, enter the appropriate range, and just take note of the IPv6 value because then we don't have to keep switching backwards and forwards to this document in this case it's 0, 5, paste in the IPv4 range, manual for IPv6, select the VPC range in the drop-down, down arrow twice, and then the left arrow and set to 0, 5, which is the unique value for the subnet, click add new subnet, this time it's app B, enter that name, availability zone B, manual for IPv4, you'll need to enter the IP range for app B, and again pay attention to the fact that the unique value for the subnet is 0, 6, manual for IPv6, select the VPC range in the drop-down, down arrow twice, right arrow until it says 0, 6, and then add new subnet again, we're going to do the last one, this time it's WebB, enter that name, availability zone B, and drop-down, manual for IPv4, copy and paste the IP range, and pay attention to 0, 7, which is the IPv6 unique value, enter the IPv4 subnet range in this box, manual for IPv6, select the VPC range in the drop-down, down arrow twice, and then right arrow until it says 0, 7, and now we've got all four subnets in AZB, so click on create subnet, and then we're going to do this one last time for availability zone C, so click on create subnet, select the animals for IPPC in the drop-down, and we're going to follow the same process, so for subnet 1 it will be SN_reserved-C, availability zone C, you'll need to enter the IPv4 range, pay attention to the IPv6 unique value, which is 0, 8, paste in that range in the box, manual for IPv6, select the VPC range, down arrow twice, set it to /64, and then right arrow until it says 0, 8, scroll down, add a new subnet, next is DVC, so enter that, availability zone C, and do the same thing as before, we'll need the IP range, and the IPv6 unique value, so 0, 9, enter that, manual for IPv6, select the VPC range in the drop-down, down arrow twice, and then click the right arrow until 0, 9 is selected, add a new subnet, then the application subnet, copy that, paste it in, availability zone C, get the IPv4 range and the unique value of IPv6, now note this is hexadecimal, so 0, 8 directly follows 0, 9, so pay attention to that, go back, paste in the IPv4 range, manual for IPv6, select the VPC range, down arrow twice to select /64, and then right arrow until you get 0, 8, then one last time, click on add new subnet, go back to the subnet document, Web C, availability zone C, get the IPv4 range and note the unique IPv6 value, paste that in, select the IPv6 range for the VPC, down arrow twice to select /64, and then right arrow all the way through to 0, B, at that point you can go ahead and click on create subnet, and that's created all four subnets in availability zone C, and all of the subnets now that are within the annuals for IPv6, at least those in AZA, AZB, and AZC, once again we're not going to create the ones in AZD which are reserved for future growth, now there's one final thing that we need to do to all of these subnets, so each of these subnets is allocated with an IPv6 range, however it's not set to auto-allocate IPv6 addressing to anything created within each of these subnets, now to do that go ahead and select SN-AP-A, click on actions, and then edit subnet settings, and I want you to check the box to say enable auto-assign IPv6 addresses, once you've done that scroll to the bottom and click on save, so that's one subnet that you've done that for, next I want you to do it for app B, follow the same process, actions, edit subnet settings, auto-assign IPv6, click on save, notice how we're not touching the IPv4 setting, we'll be changing that as appropriate later, select SN-AP-AP-A and see and again edit subnet settings, enable auto-assign IPv6, and click on save, then we're going to do the same for the database subnets, so dba, edit subnet settings, enable IPv6 and save, then dbb, check this box, save, then dbc, edit subnet settings, check this box, save, now we'll do the reserve, so reserve A, then reserve B, and then reserve C, and then finally we're going to do the web subnets, so we'll start with A, again make sure you're only changing the IPv6 box, say that, do the same with web B, and then once that's done we'll scroll down and do the final subnet, so web C, same process, IPv6, and save, and at this point you've gone through the very manual process of creating 12 subnets across three availability zones using the architecture that's shown on screen now, now in production usage in the real world you would automate this process, you wouldn't do this manually each and every time, but I think it's important that you understand how to do this process manually, so you can understand exactly what to select when configuring automation to achieve the same end goal, so whenever I'm using automation I always like to understand how it works manually, so that I can fully understand what it is that automation is doing, now at this point that is everything that I wanted to cover in this demo lesson, we're going to be continually evolving this design as we move through this section of the course, but at this point that is everything I wanted to do, so go ahead and complete this video and when you're ready, I'll look forward to you joining me in the next.

    1. Welcome to this lesson where I'm going to be talking about S3 access points, which is a feature of S3, which improves the manageability of S3 buckets, especially when you have buckets which are used by many different teams or users, or when buckets store objects with a wide range of functions.

      Now we have a good amount to cover, so let's just jump in and get started.

      S3 access points simplify the process of managing access to S3 buckets and objects.

      I want you to imagine an S3 bucket with billions of objects using many different prefixes.

      Imagine this bucket is accessed by hundreds of different teams within business.

      Now by default you would have one bucket with one really complex bucket policy.

      It would be hard to manage and prone to errors.

      Access points allow you to conceptually split this.

      You can create many access points for a bucket and each of these can have different policies, so different access controls from a permissions perspective.

      But also, each access point can be limited in terms of where they can be accessed from.

      So for example, a VPC or the internet.

      Every access point has its own endpoint address and these can be given to different teams.

      So rather than using the default endpoint for S3 and accessing the bucket as a whole, users can use a specifically created access point, along with that specific endpoint address, and get access to part of that bucket or the whole bucket, but with certain restrictions.

      Now you can create access points using either the console UI, or you can use the CLI or ABI using create-access-point.

      And it's important that you remember this command.

      Please try and remember create-access-point.

      Now it's going to be easier for you to understand this architecture if we look at it visually.

      Let's explore...

      Or how everything fits together using an architecture diagram.

      And we're going to use a typical example.

      Let's say that we have an S3 bucket and this bucket stores sensitive health information for the animals for life organization.

      Now this includes health information about the animals, but also the staff working for the business, such as medical conditions and any vaccination status.

      Now we have three different sets of staff.

      We have admin staff, field workers and vet staff.

      The admin staff look after the organization's employees.

      The field workers actually visit other countries doing wildlife studies and helping animals.

      And then the vet staff look after the medical needs of any animals which the business takes care of.

      Now in this example, the business are also using a VPC, with some instances and other resources performing data analysis functions.

      If we weren't able to use access points, then we'd have to manage the bucket as a monolithic entity, managing a large and complex bucket policy to control access for identities within this account and potentially other accounts.

      And this can become unwieldy very quickly.

      One option that we have is to use access points.

      And you can sexually think of these as mini-buckets or views on the bucket.

      So we might have three access points for our three types of users and then one for the VPC.

      Now each of these access points will have a unique DNS address for accessing it and it's this DNS address that we would give to our staff.

      But more than this, each access point also has its own policy.

      And you can think about this as functionally equivalent to a bucket policy, in that it controls access to the objects in a bucket when using that access point.

      So this can control access to certain objects, prefixes or certain tags on objects.

      And so it's super powerful.

      So now we have a unique DNS name and a unique policy.

      Essentially we have mini-buckets which are independently controlled.

      And this makes it much easier to manage the different use cases for the main bucket and our staff can access it via the access points.

      From the VPC side, access points can be set to only allow a VPC origin, which means that the access point is tied to a specific VPC.

      This will need a VPC endpoint in the VPC and the two can be tied together so that the S3 endpoint in the VPC can be configured to only allow access via the S3 access point.

      Now one really crucial thing to understand permissions-wise is that any permissions defined on an access point need to be also defined on the bucket policy.

      So in this example, if the...

      That's Stafford Grant.

      And an access via an access point policy, then the same would need to also be granted via the bucket policy.

      Now you can do delegation where on the bucket policy you grant wide open access via the access point.

      So as long as the access point is used, any action on that bucket's objects is allowed.

      And then you'll be fine, more granular control over access to objects in that bucket using the access point policies.

      And that's a pretty common permissions architecture to make things simpler to manage.

      Now I've included a link attached to this lesson with more details on permissions delegation together with some example access point policies.

      You won't need this for the exam, but if you do want to do a bit of extra reading, do make sure you check out the links included with this lesson.

      Now at this point, that's everything I wanted to cover.

      You just need to have an overview of how this works.

      At this point, that's everything.

      So thanks for watching.

      Go ahead and complete this video.

      And when you're ready, I look forward to you joining me in the next.

    1. Welcome back in this lesson, I want to cover a really important feature of S3 in Glacier that you need to be aware of for the exam.

      Now you don't need to understand the implementation just the architecture.

      S3 Select and Glacier Select are ways that you can retrieve parts of objects rather than the entire object.

      Now I expect that this will feature only in the minor way in the exam and it will be focused on architecture and features but I do want you to be aware of exactly how this feature works.

      So let's jump in and step through this architecture and what benefits it provides.

      Now you know by now that both S3 and Glacier are super scalable services.

      You can use them both to store huge quantities of data.

      S3 for example can store objects up to five terabytes in size and can store an infinite number of those objects.

      Now often when you're interacting with objects inside S3 you intentionally want to interact with that full object so you might want to retrieve that full five terabyte object.

      What's critical to understand as a solutions architect is that logically if you retrieve a five terabyte object then it takes time and it consumes that full five terabytes of transfer.

      So if you're downloading a five terabyte object from S3 into your application then you consume five terabytes of data you're accessing that full object and it takes time.

      Now you can filter this on the client side but this occurs after the figurative damage is done.

      You've already consumed that capacity, you've already downloaded that data, filtering it at the client side just means throwing away the data that you don't need.

      S3 and Glacia provide services which allow you to access partial objects so that's what's provided by the S3 select and Glacia select services and the way that you do this is that both services allow you to create a SQL like statement so cut down SQL statement.

      So you create this, you supply it to that service and then the service uses this SQL like statement to select part of that object and this part and only this part is sent to the client in a pre-filtered way so you only consume the pre-filtered part of that object, the part that you select so it's faster and it's cheaper.

      Now both S3 select and Glacia select allow you to operate on a number of file formats with this level of functionality so examples of this include comma separated values JSON can even use visa to compression comma separated values and for JSON so it's a really flexible service.

      Now visually this is how it looks.

      Now at the top we have the architecture without using S3 or Glacia select and at the bottom we have the architecture when we utilize these services so in both cases we have an application which stores its data on S3.

      So when an application interacts with S3 to retrieve any data the entire object or stream of objects are delivered to the application so the application receives everything.

      It can either accept it as a whole or it can perform its own filtering but the critical thing to understand is that any filtering performed by the application is performed inside the application.

      It doesn't impact the cost or performance.

      The only data which is filtered out is simply discarded but it's still billed for and it still takes time.

      Now contrast this to using S3 select with the same architecture so we still have the same application that interacts with the same S3 bucket.

      It places the filter point inside the S3 service itself and this allows us to use a SQL-like expression and provide this to the S3 select service.

      The S3 select service can then use the SQL-like expression and it can apply this to the raw data in S3 so in effect it's taking the raw data filtering it down but this is occurring inside the S3 service so the data that the application receives is the pre-filtered data and this means that we can achieve faster speeds and a significant reduction in cost.

      Once the filtering occurs before it's transferred to our application it means that we get substantial benefits both in speed and in cost and this is because the S3 service doesn't have to load all of the data and deliver it to the application.

      We're applying this filter at the source, the source of the data which is the S3 service itself.

      Now this is a feature which our applications will need to explicitly use but as a solutions architect it's a powerful feature that you need to understand to improve the levels of performance in any systems that you design.

      Now at this point that's everything I wanted to cover on to keep it brief because it's a product that I only expect to feature in a very very minor way in the exam and I do want you to be aware of its existence.

      So thanks for watching, go ahead, complete this video and when you're ready I look forward to you joining me in the next lesson.

    1. Welcome back.

      And in this lesson, I want to talk about S3 Object Lock.

      Now, this is something which is really important to understand for the SysOps certification, but equally, if you're developing applications running in AWS or architecting solutions for AWS, you also need to have an awareness.

      Now, we've got a fair amount to cover, so let's jump in and get started.

      S3 Object Lock is actually a group of related features which I'll talk about in this lesson, but it's something that you enable on new S3 buckets.

      If you want to turn on Object Lock for existing buckets, then you need to contact AWS support.

      And this is probably going to change over time, but at the time of creating this lesson, this is the current situation.

      Now, when you create a bucket with Object Lock enabled, versioning is also enabled on that bucket.

      Once you create a bucket with Object Lock and enable it, you cannot disable Object Lock or suspend versioning on that bucket.

      Object Lock implements a right once, read many architecture, and this is known as Worm.

      It means that you can set it so that Object Versions once created can't be overwritten or deleted.

      And just to reiterate, since this is a pretty important point, the feature requires versioning to be enabled on a bucket.

      And because of this, it's actually individual Object Versions which are locked.

      Now, when we talk about Object Lock, there are actually two ways it manages Object Retention.

      Retention periods and legal holds.

      An Object Version can have both of these, one or the other or none of them.

      And it's really important for the exam and for the real world to think of these as two different things.

      So, S3 Object Lock Retention and S3 Object Lock Legal Hold.

      Now, just like with bucket default encryption settings, these can be defined on individual Object Versions, or you can define bucket defaults for all of the Object Lock features.

      Now, this is just the feature at the high level.

      Next, I want to quickly step through the key points of the different retention methods.

      So, Retention Period and Legal Hold.

      And we're going to start with Retention Period.

      With the Retention Period style of object locking, when you create the Object Lock, you specify a Retention Period in days and/or years.

      One year means the Retention Period will end one year from when it's applied.

      Now, there are two modes of Retention Period Lock which you can apply.

      And it's really, really important that you understand how these work and the differences between the two.

      One, because it matters for the exam, and two, because if you get it wrong, it will cause a world of pain.

      The first mode is Compliance Mode.

      If you set a Retention Period on an object using Compliance Mode, it means that an Object Version cannot be deleted or overwritten for the duration of the Retention Period.

      But, it also means that the Retention Period itself cannot be reduced and the Retention Mode cannot be adjusted during the Retention Period.

      So, no changes at all to the Object Version or Retention Period settings.

      And this even includes the account root user.

      So, no identity in the account can make any changes to Object Versions, delete Object Versions, or change the Retention Settings until the Retention Period expires.

      So, this is serious business.

      This is the most strict form of Object Lock.

      Don't set this unless you really want that Object Version to stay around in its current form until the Retention Period expires.

      Now, you've used this mode as the name suggests for compliance reasons.

      An example of this might be medical or financial data.

      If you have compliance laws stating that you have to keep data, for example, for three years, with no exceptions, then this is the mode that you set.

      Now, a less strict version of this is Governance Mode.

      With this mode, you still set a Retention Period, and while active, the Object Version cannot be deleted or changed in any way.

      But you can grant special permissions to allow these to be changed.

      So, if you want a specific group of identities to be able to change settings and Object Versions, then you can provide them with the permission S3 colon Bypass Governance Retention.

      And as long as they have that permission and they provide a header along with their request, which is X-AMZ-BYPASS-GOVERNANCE-RETENTION, then they can override the Governance Mode of Retention.

      Now, an important point to understand is this last header, so X-AMZ-HIVEN, and then all the rest.

      This is actually the default for the console UI.

      And so, using the console UI, you have the S3 colon Bypass Governance Retention, you will be able to make changes to Governance Mode Retention Locks.

      So, Governance Mode is useful for a few things.

      One, if you want to prevent accidental deletion.

      Two, if you have process reasons or governance reasons to keep Object Versions.

      Or lastly, you might use it as a test of settings before picking the Compliance Mode.

      So, that's Governance Mode.

      These are both modes that can be used when using the Retention Period feature of S3 Object Locking.

      So, please make sure you understand how they both work and the differences between the two before we finish with this lesson.

      It's really, really critical that you understand.

      Now, the last overview that I want to give is S3 Object Lock Legal Hold.

      With this type, you don't actually set a Retention Period at all.

      Instead, for an Object Version, you set Legal Hold to be on or off.

      To repeat, there's no concept of retention.

      This is a binary.

      It's either on or off.

      While Legal Hold is enabled on an Object Version, you can't delete or overwrite that specific Object Version.

      An extra permission is required, which is S3 colon Put Object Legal Hold.

      And this is required if you want to add or remove the Legal Hold feature.

      And this type of Object Locking can be used to prevent accidental deletions of Object Versions, or for actual legal situations when you need to flag specific Object Versions as critical for a given case or a project.

      Now, this point, let's take a moment to summarise and look visually at how all these different features work.

      Let's start with Legal Hold.

      We start with a normal Object and we upload it to a bucket with a setting Legal Hold Status to On.

      And this means that the Object Version is locked until the Legal Hold is removed.

      In this state, the Object Version can't be deleted or changed, but you can set the Legal Hold Status to Off, at which point normal Commissions apply and the Object Version can be deleted or replaced as required.

      It's a binary.

      It's either on or off, and that isn't the concept of Retention Period.

      Next, we have the S3 Object Locks that use the Retention Period architecture.

      First, we have Governance, so we put an Object into a bucket with a Lock Configuration of Governance and specify a Retention Period.

      This creates a locked Object Version for a given number of days or years, and while it's in this state, it cannot be deleted or updated.

      With Governance Mode, this can be bypassed if you have the permissions and specify the correct header.

      And once again, this header is the default in the console, so you can adjust or remove the Object Lock or delete or replace the Object Version.

      So the important thing to realise here is while an Object is locked for a given Retention Period using the Governance Mode, you can't make any changes to Object Versions or delete them, but you can be provided with the S3 colon bypass and Governance Retention Permission, and as long as you have that and specify the X-AMZ-VIPAS-GOVERNANCE-RETENTION-TRUEHEADER, then you can override the Governance Mode Object Lock during the Retention Period.

      Then lastly, we have Compliance, which is the same architecture.

      We upload an Object.

      We specify Compliance Mode together with the Retention Period, and this creates an Object Version, which is locked for a certain period in days and years.

      The difference though is that this can't be changed.

      An Object Version can't be deleted or updated.

      The Retention Period cannot be shortened.

      The Compliance Model can't be changed to something else even by the account root user.

      This is permanent.

      Only once the Retention Period expires can the Object Version or the Retention Settings be updated.

      And for all of these, they can be set on Object Versions or what defaults can be fine.

      And that's the architecture of S3 Object Lock.

      It's critical that you understand this.

      If it takes a few watches at this lesson, then that's OK.

      Make sure you understand it in detail, including how each type differs from the others.

      And remember, they can be used in conjunction with each other, so the effects can overlap.

      You might use Legal Hold together with either Governance or Compliance, and if you do, then the effects of this could overlap, so you need to understand all of this in detail.

      But at this point, that's everything I wanted to cover in this lesson, so go ahead and complete the video.

      I mean, you ready?

      I look forward to you joining me in the next.

    1. Welcome to this mini project where you're going to get the experience of creating S3 multi-region access points.

      Now multi-region access points give you the ability to create a single S3 global endpoint and point this at multiple S3 buckets.

      It's an effective way to use a single endpoint throughout requests to the closest S3 service.

      Now in order to do this mini project you need to be logged in to an AWS account with admin permissions.

      If you're using one of my courses then you should use the IAM admin user of the general AWS account which is the management account of the organization.

      If you're not using my courses make sure you're using an identity with admin permissions.

      You'll also need to select two different AWS regions before this mini project we're going to create S3 buckets in two regions.

      I'm going to use AP, South East 2 or the Sydney region and CA Central 1 or the Canada region.

      Now the first thing to do is to move to the S3 console so type S3 in the search box at the top and then open that in a new tab.

      Once you're there we're going to create two buckets so first go ahead and click on create bucket.

      Now we'll keep the bucket naming consistent so we'll use multi-heiven region -demo -heiven and then the region that you're in.

      So in my case Sydney and then at the end I want you to append on a random number.

      In my case 1-3-3-7.

      Remember S3 bucket names need to be globally unique and this will ensure both of our buckets are.

      Once you put the name make sure you set the region correctly everything else can be left as default apart from we need to enable bucket versioning.

      So set this box under versioning to enable.

      Now scroll to the bottom and click on create bucket.

      Then we need to follow that same process again for the second bucket.

      So click on create bucket.

      Use the same bucket naming so multi-heiven region -demo -heiven and then the region name in this case Canada.

      And make sure you append your random number and set the region.

      Then scroll down, enable bucket versioning again and create the bucket.

      Now once you've got these two buckets we're going to create the multi-region access point.

      So click on multi-region access point on the menu on the left and click create multi-region access point.

      For the name you can pick whatever you want it doesn't need to be globally unique only unique within an AWS account.

      I'm going to pick really really critical cat data and then scroll down and add the buckets that you've just created.

      These can't be added or edited after creation so we need to do it now.

      Now we're going to add buckets.

      Select the two buckets and then click on add buckets to confirm.

      Once you've done that scroll down to the bottom and click create multi-region access point.

      Now this process can take worst case up to 24 hours to complete but typically it creates much faster, generally around 10 to 30 minutes.

      Now we do need this to be created before we continue.

      So go ahead and pause the video, wait for the status on this to change to ready and then you're good to continue.

      Okay so now that we've got this multi-region access point configured and it's ready to go.

      Now that we've got this multi-region access point configured we need to configure replication between the two buckets because anyone using this multi-region access point will be directed to the closest S3 bucket and we need to make sure that the data in both matches.

      So to do that go ahead and click on the multi-region access point name and go inside there and you'll see that the access point has an Amazon resource name as well as an alias.

      Now you should probably note down the Amazon resource name because we might need it later on.

      Once you've done that click on the replication and failover tab and you'll be able to see a graphical representation of any replication or failover configuration.

      If we click on the replication tab you'll see there's no replication configured.

      If we click on the failover tab you can see that we've got these two S3 buckets in different AWS regions configured as an active active failover configuration which means any requests made to this multi-region access point will be delivered to either of these S3 buckets as long as they're available.

      Now we can click on one and click on edit routing status and configure it as passive which means it will only be used if no active buckets exist.

      But in our case we want it to be active active so we'll leave both of these set to active.

      Now we want to configure replication between the buckets so we're going to scroll down to replication rules and click create replication rule.

      Now there are two templates available to start with, replicate objects amongst all specified buckets and replicate objects from one or more source buckets to one or more destination buckets.

      Now which of these you pick depends on the architecture that you're using but because we have an active active configuration we want all the buckets to be the same.

      So we're going to pick the replicate objects among all specified bucket template so this is replicating between every bucket and every other bucket.

      Essentially it creates a set of buckets which contain exactly the same data all fronted by a multi-region access point.

      So go ahead and make sure this template is selected and then click to select both of the buckets that you created.

      In my case Sydney and Canada.

      Once we've done that scroll down you can set whether you want the status to be enabled or disabled when created we're going to choose enable and you get to adjust the scope so you can either have it configured so that you can replicate objects using one or more filters or you can apply to all objects in the bucket.

      Now we want to make sure the entire bucket is replicated so we're going to use apply to all objects in the bucket.

      Now you're informed that an Ion role or roles will be generated based on your configuration and this will provide S3 with the permissions that it needs to replicate objects between the buckets.

      Now this is informational we don't need to do anything so let's move on.

      Now you're also told what encryption settings are used as well as the destination storage class so because of the template that we picked above we don't get to change the destination storage class and that's okay.

      If we scroll down to the bottom we have additional replication options, we have replication time control which applies in SLA to the replication process, we have replication metrics and notifications to provide additional rich information and we can choose whether to replicate the lead markers and whether to replicate modifications.

      Now for this mini project we're not going to use replication time control we don't need that level of SLA.

      We are going to make sure that replication metrics and notifications is selected.

      We don't want to replicate the lead markers and we do want to make sure that replica modifications sync is checked.

      So we only want replication metrics and notifications and replica modifications sync.

      So make sure that both of those are checked and then click on create replication roles.

      Now at this point all the buckets within this multi-region access point are now replicating with each other.

      In our case it's only the two, in my case it's Canada and Sydney.

      So go ahead and click on close and we can see how this graphical representation has changed showing us that we now have two-way replication in my case between Sydney and Canada.

      Now at this point we need to test out the multi-region access point and rather than having you configure your local command line interface we're going to do that with Cloud Shell.

      Now what I want you to do is to go ahead and move to a different AWS region so not the same AWS regions that either of your buckets are created in.

      What I do want you to do though is make a region close to one of your buckets.

      Now I'm going to start off with Sydney and in order to test this I'm going to switch across to the Tokyo region which is relatively close to Sydney, at least from a global perspective.

      So I'm going to click on the region drop down at the top and change it from Sydney to Tokyo.

      And what's on there I'm going to click on this icon which starts the Cloud Shell.

      If this is the first time you're using it in this region you'll probably get the welcome to AWS Cloud Shell notification.

      Just either click on close or check this box and then click on close if you don't want to see this notification again.

      Now all these commands that we're going to be running are in the instructions which are attached to this video.

      The first thing that we're going to do is to create a test file that we're going to upload to S3.

      We're going to do that using the DD command.

      So we're going to have an input of /dev/urandom which just gives us a stream of random data.

      And then for the output using the OF option we're going to create a file called test1.file.

      This is going to have a block size of 1 meg and a count of 10 which means that it's going to create a 10 meg file called test1.file.

      So run that command.

      Now once you've done that just go back to the tab that you've got open to S3.

      Scroll to the top, click on multi-region access points, check the access point that you've created and then just click on copy ARN to copy the ARN for this access point into your clipboard and then go back to Cloud Shell.

      Next I'm going to do an LS making sure I just have a file created within Cloud Shell, I do.

      And now I'm going to run this command so aws space S3, space CP for copy, space test1.file.

      So this is the local file we created within Cloud Shell and then space and then S3 colon, double forward slash and then the ARN of the multi-region access point.

      Now this command is going to copy the file that we created to this multi-region access point and this multi-region access point is going to direct us towards the closest S3 location that it serves which should be the bucket within the Sydney region.

      So go ahead and run that command.

      It's going to take a few moments to upload but when it does switch back to S3, go to buckets.

      In my case I'm going to go to the Sydney bucket and I should see the file created in this bucket.

      I do, that's good, so I'm going to go back to buckets and go to Canada and I don't yet see the object created in the Canada bucket and that's because replication can take a few minutes to replicate from the bucket where the object was stored through to the destination bucket.

      If we just give this a few moments and keep hitting refresh, after a few moments we should see the same S1.file which has been replicated from the Sydney region through to the Canada region.

      Now S3 replication isn't guaranteed to complete in a set time, especially if you haven't configured the replication time control option.

      So it's fairly normal to see a small delay between when the object gets written to one bucket and when it's replicated to another.

      Now we're going to try this with a different region, so go back to Cloud Shell and then click on the region drop down and we're going to pick a region which is close to our other bucket but not in the same region.

      So the other bucket is created in the Canada region, so I'm going to pick a close region.

      In this case I'm going to pick US East 2 which is in Ohio.

      Once I've done that I'm going to go back to Cloud Shell.

      Once I've done that I should be in Cloud Shell in a different AWS region.

      So I'll need to recreate the test file.

      In this case I'm going to call it test2.file and I'm going to use all the same options.

      And again this command is contained in the instructions attached to this video.

      So run that command and it will take a few moments to complete and then we're going to follow the same process.

      We're going to upload this file to our S3 buckets using the multi-region access point.

      So again just make sure you've got the ARN for the access point in your clipboard and then in the Cloud Shell type AWS space S3, space CP, space test2.file, space S3, colon, forward slash forward slash and then the ARN of the multi-region access point.

      Again we're going to press enter, wait for this to upload and then check the buckets.

      So run that command and then go back to S3, go to buckets.

      I'm going to go to the bucket in Canada first and I'm going to hit refresh and we should see test2.file in this bucket which is good.

      Then go to buckets and go to Sydney and we probably won't see that file just yet because it will take a few moments to replicate.

      So keep hitting refresh and eventually you should see test2.file arrives in the S3 bucket.

      Now this time we're going to run the same test but we're going to pick a region that is relatively in the middle between these two regions where our S3 buckets are created.

      In my case I'm going to change the region to APSALV1 which is the Mumbai region.

      And I'm going to follow exactly the same process and move to Cloud Shell.

      I'm going to generate a new file so in this case it's test3.file.

      Now before we upload this we're going to go to the S3 console, go to buckets and just make sure that we have a tab open for both the Canada and the Sydney bucket.

      This test is going to be uploaded from a region that's in the middle of these two buckets and so we need to be able to quickly check which bucket receives the upload directly and which bucket receives the replicated copy.

      So once you've got a tab open for both those buckets, go back to Cloud Shell and run this command.

      So aws space S3 space CP, space test3.file, space S3//// and then the name of the S3 multi-region access point.

      Once you've done that go ahead and run that command, it will take a few moments to upload and then move straight to the tabs that you've got open to the S3 buckets and refresh both of them.

      If I look at Sydney it looks as though that receives the file straight away.

      So this multi-region access point is directed us to Sydney.

      If I move to Canada and hit refresh we can see that the file's not yet arrived so this is receiving the replicated copy.

      And it does that after a few minutes we can see test3.file arrives in this bucket.

      Now this one we're going to run a test to show the type of problem that can occur if you're using this type of global replicated configuration using multi-region access points.

      What I want you to do is to open up two different Cloud Shells.

      I want one Cloud Shell open in the Canada region so the region may have one of your buckets and I want the other Cloud Shell open in the AP Southeast two region, so the Sydney region.

      So we've got one Cloud Shell open in the same region as each of our buckets.

      Now once we've done that in one region, in my case I'm going to use the Canada region, I'm going to generate a file called test4.file.

      The command is going to be the same as we've used in previous steps.

      Then I'm going to copy the ARN of the multi-region access point into my clipboard and I'm going to type this command again in the Canada Cloud Shell.

      So this command is going to copy this file to the multi-region access point.

      Now because I'm doing this in the Canada region, it's almost guaranteed that the multi-region access point is going to direct us to the bucket which is also in the Canada region.

      And so the bucket in this region is going to get the direct copy of the object and the bucket in the Sydney region is going to receive the replicated copy.

      So I'm going to fully type out this command and before I run it I'm going to move it to the Sydney region and type out a slightly different command.

      This command is also contained in the instructions attached to this lesson.

      This command is AWS space S3 space CP space and then S3 colon forward slash forward slash and then the name of the multi-region access point and then forward slash test4.file and then space and then a period.

      Now this command when we run it which we're not going to do yet is going to copy the test4.file object from the multi-region access point into our Cloud Shell.

      Remember we haven't created this object yet.

      Now because this Cloud Shell is in the Sydney region, the multi-region access point is almost certainly going to redirect us to the bucket in the Sydney region.

      So let's move back to the Canada Cloud Shell, run this command which is probably going to copy this object into the Canada bucket, then move back to the Cloud Shell in the Sydney region and then run this command to copy the object from the multi-region access point into our Cloud Shell and we receive this error.

      Now we're getting this error because the object test4.file doesn't exist within the bucket that the multi-region access point is directing us to.

      So just to reiterate what we've done, we've created this object in the Canada region using the multi-region access point which is going to have created the object in the closest S3 resource which is the bucket also in the Canada region.

      So the Canada region bucket has the direct copy of the object.

      This will then take a few minutes to replicate to the Sydney bucket because we're using this reverse copy command in the Sydney region.

      It's attempting to copy the test4.file object from the multi-region access point to our Cloud Shell but because this replication won't have occurred yet and because this is going to direct us at the bucket in the Sydney region, we get this object does not exist error.

      And this is one of the issues you can experience when you're using multi-region access points in that there's a consistency lack.

      You have to wait for the replication to occur before you can retrieve replicated objects from different buckets using the multi-region access point.

      Now this process can be improved by using the RTC option when setting up replication but this does come with additional costs.

      So this is something to keep in mind when you're using this type of architecture.

      And if we keep running this command we'll see that it keeps failing over and over again until the replication process finishes and then we can copy down the object.

      Now that is everything which I wanted to cover in this brief mini project in the multi-region access points.

      At this point all that remains is for us to clean up the account and return it to the same state as it was at the start of this mini project.

      So to do that we need to move back to the S3 console, go to multi-region access points, select the access point you created for this mini project and click delete.

      And then you need to copy and paste the name and click delete to confirm.

      Then we need to go ahead and move to buckets.

      Select each of the buckets in turn, first click on empty and you need to copy and paste or type permanently delete and then click to confirm that empty process.

      Once that's finished click on exit, select the same bucket again, this time click delete, copy and paste or type the bucket name and then click to confirm.

      And then do the same process with the other buckets so first empty it, confirm that and then delete it and confirm that.

      And then at that point all the billable elements created as part of this mini project have been deleted and we go to finish off this mini project.

      So that's everything I wanted to cover, I hope it's been useful and I hope it's given you some practical experience of how to use multi-region access points together with S3 replication.

      At this point that's everything to go ahead and complete this video and I hope you'll join me soon for another exciting mini project.

    1. Welcome back.

      In this brief demo lesson, you're going to get some experience working with S3 pre-signed URLs.

      Now, as you learned in the theory lesson, a pre-signed URL is a type of URL which can be used to grant access to certain objects within an S3 bucket where the credentials for accessing that object are encoded on the URL.

      I want to explain exactly what that means and how you can use it in the real world.

      Now, to do that, we're not going to need to create any infrastructure using CloudFormation.

      Instead, we're going to do it manually.

      So first, I want you to make sure that you're logged in to the general AWS account as the I am admin user.

      And as always, please make sure that you have the Northern Virginia region selected.

      Assuming that's all good, go ahead and type S3 in the search box at the top and open that in a new tab.

      We're going to create an S3 bucket within the general AWS account.

      So go ahead and click on create bucket to be in that process.

      Now, I want you to call the bucket animals for life media.

      And because of the unique naming requirements of S3 buckets, you'll need to add some randomness onto the end of this name.

      We'll need to select us - east - one for the region.

      We can scroll past the bucket settings for block public access.

      We don't need to enable bucket versioning.

      We won't be using any form of encryption.

      Just go ahead and click on create bucket.

      Now I want you to go inside the bucket that you've created and click on the upload button to upload an object.

      Now, at this point, we need to upload an image file to this bucket.

      Any image file will do.

      But I've included a sample one attached to this lesson if you don't have one available.

      So click the link for the image download.

      That will download an image called all5.jpeg.

      Once that's downloaded to your local machine, click on add files, select the file, open it, and then go ahead and upload the file to the S3 bucket.

      Once that's finished, you can go ahead and click on close.

      And you'll see now that we've got our S3 bucket and one object uploaded to that bucket called all5.jpeg.

      Next, go ahead and click on that object.

      Now, I want to demonstrate how you can interact with this object in a number of different ways.

      And the detail really matters here.

      So we really need to be sure of exactly what differences there are between these different methods.

      The first thing I want you to do is towards the top right of the screen, click on the open button.

      You might get a pop-up notification if you do just allow pop-ups.

      And what that's going to do is open a jpeg object in a new tab.

      Now, I want to point out a number of really important points about how this object has been opened.

      If you take a moment to review the URL that's been used to open this object, you'll note that a number of pieces of information have been specified on the URL, including AMZ-SecurityToken.

      So essentially, a form of authentication has been provided on the URL which allows you to access this object.

      Now, I want you to contrast this by what happens if we go back to the S3 bucket and just copy down the object URL into our clipboard and note how this does not contain any additional authentication information.

      It's just the raw URL for this object.

      So copy that into your clipboard and then open a new tab and paste that in.

      What you're going to see is an access denied message.

      And this makes sense because we're now attempting to access this object as an unauthenticated identity, just like any internet user, anyone browsing to this bucket would be doing.

      We're not providing any authentication.

      And so the only way that we can access the object in this way by not providing any authentication is if we made the bucket public and the bucket currently isn't public, which is why we're getting this access denied message.

      Just to reiterate, we can access it by using the previous method because by opening it from within the console, the console is intelligent enough to add authentication information onto the URL which allows us to access this object.

      So those are the differences between these two methods.

      One is providing authentication and the other isn't.

      So right now the only entity that is able to access any of the objects within this bucket is this AWS account and specifically the I am admin user.

      The object has no public access.

      So now what we're going to do is we're going to work with the scenario that you want to grant access to this object to somebody else for a limited amount of time.

      So you don't want to provide the URL that includes authentication information, you want to provide a URL which allows access to that object for a limited amount of time.

      That's important.

      So we're going to generate a time limited pre-sign URL.

      So go ahead and click on the Cloud Shell icon and this is going to open a Cloud Shell using the identity that you're currently logged in at.

      So it's going to open a shell much like the one you'll see when you're connected to an EC2 instance but the credentials that you have in the shell are going to be the credentials of the identity that you're currently logged into AWS using in our case the I am admin user with a general AWS account.

      You'll see a message saying preparing your terminal and then you'll be logged in to what looks like an EC2 instance prompt.

      You can use the AWS CLI tool so I'm able to run an AWS space S3 space LS and this shell will interact with your current AWS account using your current credentials.

      So in my case I'm able to see the Animals for Life media bucket which is in my general AWS account because I'm currently logged into this Cloud Shell using the credentials of my I am admin user.

      Now to generate a pre-sign URL we have to use this command.

      So use AWS to use the command line tools and then a space S3 because we're using the S3 service and then a space and then the word pre-sign because we want to generate a pre-signed URL and then a space and then we need the S3 URI to this object.

      Now we can get that from the S3 console.

      For every object you can see this unique URI so go ahead and copy this into your click mode go back to the Cloud Shell paste it in and then we'll need space and then we'll use double hyphen expires hyphen in and then a space and then we need to provide the number of seconds that this pre-signed URL will be valid for.

      In this case we're going to use 180 which is a total of three minutes so this is in seconds so three minutes is 180 seconds.

      So go ahead and press enter and this will generate you a unique pre-signed URL.

      So go ahead and copy this into your click board and it is a really long URL so make sure that you get everything including HTTPS all the way to the end of this URL.

      Copy that into your click board and then I want you to open a new tab and load that URL and there you go you can see this object loads up using this pre-signed URL.

      Now this URL is valid only for 180 seconds.

      To demonstrate that I'm going to skip ahead 180 seconds and demonstrate exactly what happens when this URL expires.

      After 180 seconds when a next refresh we see this access denied page with a message request has expired.

      So you can see how pre-signed URLs are a really effective way of granting access to objects within an S3 bucket for a limited amount of time.

      Now there are a number of really interesting aspects to pre-signed URLs that you really need to understand as an architect, a developer or an engineer and I want to go through these really interesting aspects of pre-signed URLs before we finish up with this demo lesson.

      Now first just to make everything easier to see I'm going to close down any of these tabs that we got open to this all five object.

      I'm going to go back to the cloud shell and I'm going to generate a new pre-signed URL but this time I'm going to use a much larger expires in time.

      So I'm going to press the up arrow to return to the previous command.

      I'm going to delete this 180 second expiring time and instead I'm going to use 604,800 which is a pretty high number but this is something that we can pick so that we won't have any unintentional expires at the URL as part of this demo lesson.

      So pick something crazily large just to make sure that it doesn't expire until we're ready.

      So we're generating another URL and so this is an additional unique pre-signed URL.

      So I'm going to select all of this URL and you need to do the same go ahead and copy that into your clipboard and open it in a new tab.

      Now we can see that that pre-signed URL has opened.

      Keep in mind that we've generated this using the identity that we're currently logged into AWS using.

      So next what I'm going to do is move back to the AWS console and I'm going to click on the services drop down and move across to IAM.

      So I'm going to open IAM up in a new tab.

      I'm going to select the users option, select the IAM admin user, I'm going to add an inline policy to this user, select JSON.

      Now attached to this lesson is another link for a file called deny s3.json.

      Go ahead and click that link and then you'll need to get the contents of the file.

      So copy all the contents of the file into your clipboard and then select all of this JSON and paste in the contents of that file and this is an explicit deny policy which denies our user so IAM admin any access to s3.

      So this essentially prevents our IAM admin user any level of access to s3.

      Go ahead and click on review policy for name, call it deny s3 and click on create policy and this will attach this as an inline policy to our IAM admin user.

      Now I'm going to clear the screen within this cloud shell to make it easier to see and then I'm going to run an AWS space s3 space ls and press enter.

      Now we get an access deny when accessing the s3 service because we just added an explicit deny onto the IAM admin user and remember the rule that applies to permissions deny allow deny and explicit deny always overrules everything else and so even though the IAM admin user has administrative permissions by applying this deny s3 policy which has an s3 explicit deny this wins so currently we have no access to s3.

      Now let's return to the tab where we have this pre-signed url open.

      So remember this pre-signed url was generated at the time when we had access to s3 so now let's refresh this page and now we get an access denied message and this is one of those interesting aspects of pre-signed urls.

      When you generate a pre-signed url using an identity such as an IAM user that url has your access permissions so you generate a pre-signed url and anyone using that pre-signed url for the duration that it's active will be interacting with that one object on s3 as though they were using your identity.

      Now if you adjust the permissions on your identity as we've just done by denying access to s3 it means that that pre-signed url will also have that deny s3 permissions and so this pre-signed url now no longer has any access to anything in s3 including the object that it was configured to provide access to so now let's go back to our cloud shell and what we're going to do now remember we are still denied access to s3.

      Let's press the app arrow and move back to the command which we used to generate this pre-signed url.

      Now let's regenerate another pre-signed url.

      Now note how even though we are denied access to any part of s3 we can generate a pre-signed url which points at this one specific object all five dot jpeg so we're not prevented from generating a pre-signed url for something that we have no access to.

      Now if we copy this into our clipboard move to the tab that we already have open and just replace this url with the pre-signed url that you just generated.

      Note that we're still denied we're still denied because although we could generate this pre-signed url we've generated it using an identity which has no access to s3 so the next interesting fact to pre-signed urls that I want to demonstrate if we go back to the iam console and now we remove this deny s3 policy from our iam admin user now our iam admin user once again has access to s3 so if we go back to the cloud shell we can run natal us space s3 space ls and press enter and now we can access s3 again from this cloud shell remember the cloud shell is using commissions based on our iam admin user now let's go back to the pre-signed url and now if we refresh this now we can access this object again so I'm doing this to illustrate that when you generate a pre-signed url that pre-signed url is linked to the identity that generates it whatever the permissions are of that identity when the pre-signed url is used that is the permissions that that pre-signed url has so for the duration it always has the same permissions that the identity which generated it has at that very moment so that's something that you need to keep in mind now more interestingly is that you can actually generate a pre-signed url for a non-existent object there's nothing preventing that from occurring so if we go back to the cloud shell we press the up arrow and a number of times to bring up this pre-signed url command and this time we try to generate a pre-signed url for an object which doesn't exist so if I change this all 5 to all 1 3 3 7 and press enter it will generate a pre-signed url that pre-signed url will be valid to access an object called all 1 3 3 7 .jpeg inside this bucket but because there's no such object in that bucket if I try to use it I won't be able to do so so if I open that new invalid pre-signed url I'll get the message that the specified key does not exist but we can generate pre-signed urls for non-existent objects now one more interesting thing about pre-signed urls which I'm not going to demonstrate is if you generate a pre-signed url using temporary credentials you get by issuing a roll so for example if we logged into an ec2 instance which had an instance roll on that instance and then we generated a pre-signed url even if we set a huge expiry time so 604800 that pre-signed url would stop working when those temporary credentials for that roll also stopped working now it is possible to generate a pre-signed url from the console UI this is a relatively recent change from the object if you click on the object actions drop down you can click share with the pre-signed url you have to set the same settings so what you want the expiry to be in this particular case let's say 60 minutes and then I can go ahead and click on create pre-signed url now that's automatically copied into my clipboard and I can go ahead and move to a different tab paste that in and we can open the console UI generated pre-signed url so that's just an alternative way of doing the same process that we just used the cloud shell for now that's everything that I wanted to demonstrate in this demo lesson about pre-signed url and don't worry we're going to be talking about these more later in the course as well as looking at some alternatives what we need to do to tidy up this lesson is go back to the AWS console move to the S3 console and then just go ahead and empty and delete the bucket that you created so select animals for live media click on empty type or paste in permanently delete and then confirm or once that's successfully empty click on exit the bucket should still be selected and then go ahead and click on delete delete confirm that with the name of the bucket and then go ahead and delete the bucket and at this point we've cleaned up the account and the resources back in the same state as they were at the start of the lesson so I hope you've enjoyed this brief demo lesson go ahead complete the video and when you're ready I look forward to you joining me in the next video.

    1. Welcome back and in this lesson I want to talk to you about S3 pre-signed URLs.

      Pre-signed URLs are a way that you can give another person or application access to an object inside an S3 bucket using your credentials in a safe and secure way.

      Let's have a look at how this works architecturally.

      To illustrate how pre-signed URLs work, let's use this example architecture, an S3 bucket which doesn't have any public access configured.

      So it's still in a default private configuration.

      This means that in order to access the bucket or any resources in it an IAM user such as IAM admin would have to authenticate to AWS and be authorized to access the resource.

      IAM admin would send credentials along with the access request, AWS would validate them at the time that the request is made and only then grant access to the object in S3.

      Now one issue that we have is that because the bucket is private only authenticated users are able to access it.

      Our masked man here has no way of providing authentication information to AWS because he doesn't have any and so any request that's unauthenticated would fail.

      Now if giving the mystery user access to S3 is an essential requirement for the business then there are three common solutions at this point and none of them are ideal.

      Number one is to give the mystery user an AWS identity.

      Number two is to give the mystery user some AWS credentials to use or number three is to make the bucket or the object public.

      Now none of these are ideal.

      If the user only needs short term access to the bucket and objects the effort of supplying an identity seems excessive.

      Just giving some credentials to the user appears on the surface to be a security risk and definitely if it's somebody else's credentials that's just bad practice.

      Making the bucket public for anything which has sensitive data in it also appears to be less than ideal.

      So one solution that AWS offers is to use pre-signed URLs and let's look at how this would work with an architecture example.

      IAM admin is an AWS identity with some access rights granted via a permissions policy.

      So IAM admin can make the request to S3 to generate a pre-signed URL.

      She would need to provide her security credentials, specify a bucket name, an object key and an expiry date and time as well as indicate how the object would be accessed.

      And S3 will create a pre-signed URL and return it.

      This URL will have encoded inside it the details that IAM admin provided.

      So which bucket is for which object is for it will be encoded with the fact that the IAM admin user generated it and it will be configured to expire at a certain date and time as requested by the IAM admin user.

      The URL could then be passed to our mystery user and he or she could use it to access a specific object in the specific S3 bucket up until the point at which it expires.

      When the pre-signed URL is used the holder of that URL is actually interacting with S3 as the person who generated it.

      So for that specific object in that specific bucket until the timer expires our masked man is actually IAM admin.

      Pre-signed URLs can be used for both downloads from S3 so get operations or uploads to S3 known as put operations.

      So this type of architecture might be useful for the animals for life remote workers if they don't have access to AWS accounts or accessing from a secure location and just need to upload one specific object.

      Now there's another type of architecture which commonly uses pre-signed URLs and I've mentioned this earlier in the course when I was talking about some benefits of S3.

      I want you to think about a traditional application architecture.

      On the left we have an application user with a laptop.

      We've got an application server in the public cloud in the middle which hosts the application and for example say let's say this is a video processing application for wildlife videos that the animals for life organization manages.

      We've learned that one of the real strengths of S3 is its ability to host the large media files and so the large wildlife video files have been migrated from the application server to a media S3 bucket.

      But by doing this we've introduced a problem.

      Previously the videos were hosted on the application server and it could control access to the video files.

      If we host them on an S3 bucket then either every user needs an AWS identity so an IAM user to access the videos or we need to make the videos public so that the user running the web application can download them into her browser and neither of those are ideal.

      Luckily pre-signed URLs offer a solution.

      With pre-signed URLs we can keep the bucket private.

      Then we can create an IAM user in our AWS account for this application.

      Remember when I talked about choosing between an IAM user or a role I said that if you could visualize how many of a certain thing that would be using an identity then it's likely to suit an IAM user.

      Well in this case we have one application and application service accounts are a common scenario where IAM users are used.

      So for this example we create an IAM user for the application.

      So when our application user interacts with a web application she makes a request and that's step one.

      The request might be for a page of information on a bushfire which is currently happening in Australia which has a video file associated with it.

      The application that's running on the server knows that it can directly return the information that the request is asking for but that the video is hosted on the private S3 bucket.

      So it initiates a request to S3 asking for it to generate a pre-signed URL for that particular video object using the permissions of the IAM user that the application is using.

      So IAM app one.

      The S3 service then creates a URL which has encoded within it the authentication information for the IAM app one user.

      The access to one object in the bucket for a short time limited basis maybe two to three hours.

      And the S3 service then returns that to the application server and through to the end user.

      The web application that's running on the user's laptop then uses this pre-signed URL to securely access the particular object that's stored on the media bucket.

      Pre-signed URLs are often used when you offload media into S3 or as part of serverless architectures where access to a private S3 bucket needs to be controlled and you don't want to run thick application servers to broke at that access.

      And we'll look at serverless architectures later in the course.

      For now I just want to be sure that you understand what pre-signed URLs do and in the next lesson which is a demo lesson you're going to get the chance to experiment yourself and generate a pre-signed URL.

      Pre-signed URLs can be used to access an object in a private S3 bucket with the access rights of the identity which generates them.

      They're time limited and they encode all of the authentication information needed inside.

      And they can be used to upload objects and download objects from S3.

      Now I want to show you this in a demo because it's far easier to actually do it.

      But before we do that I want to step through some exam power-ups.

      There are a couple of really interesting facts about generating pre-signed URLs which might help you out in some exam questions.

      Now first and this is a fairly odd behavior but you can create a pre-signed URL for an object that you have no access to.

      The only requirement for generating a pre-signed URL is that you specify a particular object and an expiry date and time.

      If you don't have access to that object you can still generate a pre-signed URL which also because it's linked to you will have no access to that object.

      So there aren't many use cases where this is applicable but you do need to be aware that it is possible to generate a pre-signed URL when you don't have any access.

      Now when you're using the URL, so when you utilize a URL and attempt to access an object, the permissions that you have access to match the identity that generated it.

      And it's important to understand that it matches the permissions that the identity has right now.

      So at the point when you use the URL, the URL has the same permissions as the identity that generated it has right now.

      So if you get an access denied error when you attempt to use a pre-signed URL to access an object, it could be that the identity that generated that URL never had access or it could be that it simply doesn't have access right now.

      And they're two very important nuances to understand about pre-signed URLs.

      So when you're using a URL to access an object, it matches the current permissions of the identity that generated that URL.

      Now that's fairly okay for an IM user.

      As I demonstrated on the previous example of the application, generally you would create an IM user for the application and this IM user would have fairly static permissions.

      So the permissions generally wouldn't change between when you created the URL and when your customer is using that URL within an application.

      But don't generate pre-signed URLs based on an IM role.

      You can in theory assume an IM role, which remember gives you temporary credentials, and then you can use those temporary credentials to generate a pre-signed URL.

      A pre-signed URL can have a much longer validity period than those temporary credentials.

      So those temporary credentials will generally expire well before a pre-signed URL does.

      So if you generate a pre-signed URL using a role and then those temporary credentials expire, those credentials are no longer valid.

      And so that URL will stop working.

      And so it's almost never a good idea to generate a pre-signed URL using an IM role.

      You should always use long-term identities.

      So generally an IM user.

      With that being said though, that is everything that I wanted to cover in this theory lesson.

      So go ahead, complete the video, and when you're ready, I'll look forward to you joining me in the next.

    1. So this fall, he plans to scrap many of his writing assignments, including the experiential-learning one that was once so meaningful to many of his students. “Because of those people at the bottom of the scale making it impossible for me to do my work,” he says of AI users, “all those people at the upper end of the scale will never have that good experience.” Some of those better students might even have chosen to become religious-studies majors.

      Scrapping nearly all the writing assignments takes away so much of the chances to learn, experiment and improve on their skills. Especially when many other professors aren't going to follow this example, either.

      This does completely shut down usage of Generative AI, but at the cost of an effective class. Whether it's the fault of AI usage or the teacher is up to anyone, but I'm not on the teacher's side on this.

    2. Administrators who felt positively about AI focused on the need to prepare students for an AI-infused workplace, and said that it could spur new ways of thinking about problems and enhance learning through tools such as AI tutors.

      While I understand where the administrators are coming from, it's important to keep accessibility in mind. Let me provide an example on this:

      You're teaching someone to drive in your car, and your car has cameras that covers blind spots and the rear view. Do you make the student check their blind spots manually? If not, how will they know to do so in a car without cameras?

      I believe you should teach students to develop these skills without AI's help. If you don't, how will they function without AI?

    1. Author response:

      The following is the authors’ response to the original reviews.

      Reviewer #1 (Public Review):

      Weakness 1. Enhancing Reproducibility and Robustness: To enhance the reproducibility and robustness of the findings, it would be valuable for the authors to provide specific numbers of animals used in each experiment. Explicitly stating the penetrance of the rod-like neurocranial shape in dact1/2-/- animals would provide a clearer understanding of the consistency of this phenotype. 

      In Fig. 3 and Fig. 4 animal numbers were added to the figure and figure legend (line 1111). In Fig. 5 animal numbers were added to the figure. We now state that dact1/2-/- animals exhibit the rod-like neurocranial shape that is completely penetrant (Line 260). 

      Weakness 2. Strengthening Single-Cell Data Interpretation: To further validate the single-cell data and strengthen the interpretation of the gene expression patterns, I recommend the following: 

      -Provide a more thorough explanation of the rationale for comparing dact1/2 double mutants with gpc4 mutants.

      -Employ genotyping techniques after embryo collection to ensure the accuracy of animal selection based on phenotype and address the potential for contamination of wild-type "delayed" animals.

      -Supplement the single-cell data with secondary validation using RNA in situ or immunohistochemistry techniques. 

      An explanation of our rationale was added to the results section (Lines 391403) and a summary schematic was added to Figure 6 (panel A).

      Genotyping of the embryos was not possible but quality control analysis by considering the top 2000 most variable genes across the dataset showed good clustering by genotype, indicating the reproducibility of individuals in each group (See Supplemental Fig. 4).

      The gene expression profiles obtained in our single-cell data analysis for gpc4, dact1, and dact2 correlate closely with our in situ hybridization analyses. Further, our data is consistent with published zebrafish single-cell data. We validated our finding of increased capn8 expression in dact1/2 mutants by in situ hybridization. Therefore we are confident in the robustness of our single-cell data.  

      Weakness 3. Directly Investigating Non-Cell-Autonomous Effects: To directly assess the proposed non-cell-autonomous role of dact1/2, I suggest conducting transplantation experiments to examine the ability of ectodermal/neural crest cells from dact1/2 double mutants to form wild-type-like neurocranium.  

      The reviewer’s suggestion is an excellent experiment and something to consider for future work. Cell transplant experiments between animals of specific genotypes are challenging and require large numbers. It is not possible to determine the genotype of the donor and recipient embryos at the early timepoint of 1,000 cell stage where the transplants would have to be done in the zebrafish. So that each transplant will have to be carried out blind to genotype from a dact1+/-; dact2+/- or dact1-/-; dact2+/- intercross and then both animals have to be genotyped at a subsequent time point, and the phenotype of the transplant recipient be analyzed. While possible, this is a monumental undertaking and beyond the scope of the current study.

      Weakness 4. Further Elucidating Calpain 8's Role: To strengthen the evidence supporting the critical role of Calpain 8, I recommend conducting overexpression experiments using a sensitized background to enhance the statistical significance of the findings. 

      We thank the reviewer for their suggestion and have now performed capn8 overexpression experiments in embryos generated from dact1/2 double heterozygous breeding. We found a statistically significant effect of capn8 overexpression in the dact1+/-,dact2+/- fish (Lines 462-464 and Fig. 8C,D). 

      Minor Comments:  

      Comment: Creating the manuscript without numbered pages, lines, or figures makes orientation and referencing harder.  

      Revised

      Comment: Authors are inconsistent in the use of font and adverbs, which requires extra effort from the reader. ("wntIIf2 vs wnt11f2 vs wnt11f2l"; "dact1/2-/- vs dact1/dact2 -/-"; "whole-mount vs wholemount vs whole mount").  

      Revised throughout.

      Comment: Multiple sentences in the "Results" belong to the "Materials and Methods" or the "Discussion" section. 

      We have worked to ensure that sentences are within the appropriate sections of the manuscript.

      Comment: Abstract:

      "wnt11f2l" should be "wnt11f2"  

      Revised (Line 24).

      Comment: Main text:

      Page 5 - citation Waxman, Hocking et al. 2004 is used 3x without interruption any other citation. 

      Revised (Line 112).

      Page 9 - "dsh" mutant is mentioned once in the whole manuscript - is this a mistake?

      Revised, Rewritten (Line 196).

      Page 10 - Fig 2B does not show ISH.

      Revised (Line 229).

      Page 11 - "kyn" mutant is mentioned here for the first time but defined on page 15.

      Revised (Line 245). Now first described on page 4.

      Page 14 - "cranial CNN" should be CNCC.

      Revised. (Line 334)

      Page 16 - dact1/dact2/gpc4: Fig. 5C is used but it should be Fig 5E.

      Revised. (Line 381)

      Page 18 - dact1/2-/- or dact1-/-, dact2-/-. 

      Revised. (Line 428)

      Comment: Methods:

      Page 24 - ZIRC () "dot" is missing. ChopChop ")" is missing. "located near the 5' end of the gene" - In the Supplementary Figure 1 looks like in the middle of the gene.

      Revised. (Lines 600, 609, 611, respectively).

      Page 25 - WISH -not used in the main text.

      Revised. (Line 346).

      Page 26 - 4% (v/v) formaldehyde; at 4C - 4{degree sign}C; 50% (v/v) ethanol; 3% (w/v) methylcellulose.

      Revised. (Lines 659, 660, 662).

      Page 27 - 0.1% (w/v) BSA. 

      Revised. (Line 668).

      Comment: Discussion:

      The overall discussion requires more references and additional hypotheses. On page 20, when mentioning 'as single mutants develop normally,' does this refer to the entire animals or solely the craniofacial domain? Are these mutants viable? If they are, it's crucial to discuss this phenomenon in relation to prior morpholino studies and genetic compensation.

      Observing how the authors interpret previously documented changes in nodal and shh signaling would be beneficial. While Smad1 is discussed, what about other downstream genes? Is shh signaling altered in the dact1/2 double mutants? 

      We have revised the Discussion to include more references (Lines 473, 476, 483, 488, 491, 499, 501, 502, 510, 515, 529, 557, 558) and additional hypotheses (Lines 503-505, 511-519, 522-525). We have added more specific information regarding the single mutants (Lines 270-275, 480-493, Fig. S3). We have added discussion of other downstream genes, including smad1 (Lines 561-572) and shh (Lines 572-580).

      Comment: Figures:

      Appreciating differences between specimens when eyes were or were not removed is quite hard.

      Yes this was an unfortunate oversight, however, the key phenotype is the EP shown in the dissections.

      Fig 1. - wntIIf2 vs wnt11f2? C - Thisse 2001 - correct is Thisse et al. 2001.

      Revised typo in Fig 1. (And Line 1083).

      Fig 1E: These plots are hard to understand without previous and detailed knowledge. Authors should include at least some demarcations for the cephalic mesoderm, neural ectoderm, mesenchyme, and muscle. Missing color code.

      We have moved this data to supplementary figure S1 and have added labels of the relevant cell types and have added the color code.

      Comment:- Fig 2 - In the legend for C - "wildtype and dact2-/- mutant" and "dact1/2 mutant"; in the picture is dact1-/-, dact2-/-.

      Revised (Line 1105).

      Fig 2 - B - it is a mistake in 6th condition dact1: 2x +/+, heterozygote (+/-) is missing.

      Revised Figure 2B.

      Fig 4. - Typo in the legend: dact1/"t"2-/- .

      Revised. (Line 1127).

      Fig 8C - In my view, when the condition gfp mRNA says "0/197, " none of the animals show this phenotype. I assume the authors wanted to say that all the animals show this phenotype; therefore, "197/197" should be used.

      We have removed this data from the figure as there were concerns by the reviewers regarding reproducibility. 

      Fig S1 - Missing legend for the 28 + 250, 380 + 387 peaks? RT-qPCR - is not mentioned in the Materials and Methods. In D - ratio of 25% (legend), but 35% (graph).

      Revised.(Line 1203, Line 625, Line 1213, respectively).

      Fig S2 - The word "identified" - 2x in one sentence. 

      Revised. (Line 1230).

      Reviewer #2 (Public Review):

      Weakness(1) While the qualitative data show altered morphologies in each mutant, quantifications of these phenotypes are lacking in several instances, making it difficult to gauge reproducibility and penetrance, as well as to assess the novel ANC forms described in certain mutants.  

      In Fig. 3 and Fig. 4 animal numbers were added to the figure legend. In Fig. 5 animal numbers were added to the figure to demonstrate reproducibility. We now state that dact1/2-/- animals exhibit the rod-like neurocranial shape that is completely penetrant (Line 260). As the altered morphologies that we report are qualitatively significant from wildtype we did not find it necessary to make quantitative measurements. For experiments in which it was necessary to in-cross triple heterozygotes (Fig 3, Fig. 5), we dissected and visually analyzed the ANC of at least 3 compound mutant individuals. At least one individual was dissected for the previously published or described genotypes/phenotypes (i.e. wt, wntllf2-/-, dact1/2-/-, gpc4-/-, wls/-). We realize quantitative measurements may identify subtle differences between genotypes. However, the sheer number of embryos needed to generate these relatively rare combinatorial genotypes and the amount of genotyping required prevented quantitative analyses. 

      Weakness 2) Germline mutations limit the authors' ability to study a gene's spatiotemporal functional requirement. They therefore cannot concretely attribute nor separate early-stage phenotypes (during gastrulation) to/from late-stage phenotypes (ANC morphological changes). 

      We agree that we cannot concretely attribute nor separate early and latestage phenotypes. Conditional mutants to provide temporal or cell-specific analysis are beyond the scope of this work. Here we speculate based on evidence obtained by comparing and contrasting embryos with grossly similar early phenotypes and divergent late-stage phenotypes. We believe our findings contribute to the existing body of literature on zebrafish mutants with both early convergent extension defects and craniofacial abnormalities.   

      Weakness (3) Given that dact1/2 can regulate both canonical and non-canonical wnt signaling, this study did not specifically test which of these pathways is altered in the dact1/2 mutants, and it is currently unclear whether disrupted canonical wnt signaling contributes to the craniofacial phenotypes, even though these phenotypes are typical non-canonical wnt phenotypes. 

      Previous literature has attributed canonical wnt, non-canonical wnt, and nonwnt functions to dact, and each of these likely contributes to the dact mutant phenotype (Lines 87-89). We performed cursory analyses of tcf/lef:gfp expression in the dact mutants and did not find evidence to support further analysis of canonical wnt signaling in these fish. Single-cell RNAseq did not identify differential expression of any canonical or non-canonical wnt genes in the dact1/2 mutants.

      Further research is needed to parse out the intracellular roles of dact1 and dact2 in response to wnt and tgf-beta signaling. Here we find that dact may also have a role in calcium signaling, and further experiments are needed to elaborate this role.      

      Weakness (4) The use of single-cell RNA sequencing unveiled genes and processes that are uniquely altered in the dact1/2 mutants, but not in the gpc4 mutants during gastrulation. However, how these changes lead to the manifested ANC phenotype later during craniofacial development remains unclear. The authors showed that calpain 8 is significantly upregulated in the mutant, but the fact that only 1 out of 142 calpainoverexpressing animals phenocopied dact1/2 mutants indicates the complexity of the system. 

      To further test whether capn8 overexpression may contribute to the ANC phenotype we performed overexpression experiments in the resultant embryos of dact1/dact2 double het incross. We found the addition of capn8 caused a small but statistically significant occurrence of the mutant phenotype in dact1/2 double heterozygotes (Fig.8D). We agree with the reviewer that our results indicate a complex system of dysregulation that leads to the mutant phenotype. We hypothesize that a combination of gene dysregulation may be required to recapitulate the mutant ANC phenotype. Further, as capn8 activity is regulated by calcium levels, overexpression of the mRNA alone likely has a small effect on the manifestation of the phenotype. 

      Weakness (5) Craniofacial phenotypes observed in this study are attributed to convergent extension defects but convergent extension cell movement itself was not directly examined, leaving open if changes in other cellular processes, such as cell differentiation, proliferation, or oriented division, could cause distinct phenotypes between different mutants. 

      Although convergent extension cell movements were not directly examined, our phenotypic analyses of the dact1/2 mutant are consistent with previous literature where axis extension anomalies were attributed to defects in convergent extension (Waxman 2004, Xing 2018, Topczewski 2001). We do not attribute the axis defect to differentiation differences as in situ analyses of established cell type markers show the existence of these cells, only displaced relative to wildtype (Figure 1). We agree that we cannot rule out a role for differences in apoptosis or proliferation however, we did not detect transcriptional differences in dact1/2 mutants that would indicate this in the single-cell RNAseq dataset. Defects in directed division are possible, but alone would not explain that dact1/2 mutant phenotype, particularly the widened dorsal axis (Figure 1).

      Major comments:  

      Comment (1) The author examined and showed convergent extension phenotype (CE) during body axis elongation in dact1/dact2-/- homozygous mutants. Given that dact2-/- single mutants also displayed shortened axis, the authors should either explain why they didn't analyze CE in dact2-/- (perhaps because that has been looked at in previously published dact2 morphants?) or additionally show whether CE phenotypes are present in dact1 and dact2 single mutants.  

      The authors should quantify the CE phenotype in both dact2-/- single mutants and dact1/dact2-/- double mutants, and examine whether the CE phenotypes are exacerbated in the double mutants, which may lend support to the authors' idea that dact1 can contribute to CE. The authors stated in the discussion that they "posit that dact1 expression in the mesoderm is required for dorsal CE during gastrulation through its role in noncanonical Wnt/PCP signaling". However, no evidence was presented in the paper to show that dact1 influences CE during body axis elongation.  

      Because any axis shortening in shortening in dact2-/- single mutants was overcome during the course of development and at 5 dpf there was no noticeable phenotype, we did not analyze the single mutants further.  

      We have added data to demonstrate the resulting phenotype of each combinatorial genotype to provide a more clear and detailed description of the single and compound mutants (Fig. S3). 

      Our hypothesis that dact1 may contribute to convergent extension is based on its apparent ability to compensate (either directly or indirectly) for dact2 loss in the dact2-/- single mutant. 

      Comment (2) Except in Fig. 2, I could not find n numbers given in other experiments. It is therefore unclear if these mutant phenotypes were fully or partially penetrant. In general, there is also a lack of quantifications to help support the qualitative results. For example, in Fig. 4, n numbers should be given and cell movements and/or contributions to the ANC should be quantified to statistically demonstrate that the second stream of CNCC failed to contribute to the ANC.  

      Similarly, while the fan-shaped and the rod-shaped ANCs are very distinct, the various rod-shaped ANCs need to be quantified (e.g. morphometry or measurements of morphological features) in order for the authors to claim that these are "novel ANC forms", such as in the dact1/2-/-, gpc4/dact1/2-/-, and wls/dact1/2-/- mutants (Fig. 5).  

      We have added n numbers for each experiment and stated that the rod-like phenotype of the dact1/2-/- mutant was fully penetrant. 

      Regarding CNCC experiments, we repeated the analysis on 3 individual controls and mutants and did not find evidence that CNCC migration was directly affected in the dact1/2 mutant. Rather, differences in ANC development are likely secondary to defects in floor plate and eye field morphometry. Therefore we did not do any further analyses of the CNCCs.

      Regarding figure 5, we have added n numbers. We dissected and analyzed a minimum of three triple mutants (dact1/2-/-,gpc4-/- and dact1/2-/-,wls-/-) and numerous dact1/s double mutants and found that the triple mutant ANC phenotype was consistent and recognizably different enough from the dact1/2-/-, or gpc4 or wls single mutant that morphometry measurements were not needed. Further, the triple mutant phenotype (narrow and shortened) appears to be a simple combination of dact1/2 (narrow) and gpc4/wls (shortened) phenotypes. As we did not find evidence of genetic epistasis, we did not analyze the novel ANC forms further.

      Comment (3): The authors have attributed the ANC phenotypes in dact1/2-/- to CE defects and altered noncanonical wnt signaling. However, no evidence was presented to support either. The authors can perhaps utilize diI labelling, photoconversionmediated lineage tracing, or live imaging to study cell movement in the ANC and compare that with the cell movement change in the gpc4-/- , and gpc4/dact1/2-/- mutants in order to first establish that dact1/2 affect CE and then examine how dact1/2 mutations can modulate the CE phenotypes in gpc4-/- mutants.  

      Concurrently, given that dact1 and dact2 can affect (perhaps differentially) both canonical and non-canonical wnt signaling, the authors are encouraged to also test whether canonical wnt signaling is affected in the ANC or surrounding tissues, or at minimum, discuss the potential role/contribution of canonical wnt signaling in this context.  

      Given the substantial body of research on the role of noncanonical wnt signaling and planar cell polarity pathway on convergent extension during axis formation (reviewed by Yang and Mlodzik 2015, Roszko et al., 2009) and the resulting phenotypes of various zebrafish mutants (i.e. Xing 2018, Topczewski 2001), including previous research on dact1 and 2 morphants (Waxman 2004), we did not find it necessary to analyze CE cell movements directly.  

      Our finding that CNCC migration was not defective in the dact1/2 mutants and the knowledge that various zebrafish mutants with anterior patterning defects (slb, smo, cyc) have a similar craniofacial abnormality led us to conclude that the rod-like ANC in the dact1/2 mutant was secondary to an early patterning defect (abnormal eye field morphology). Therefore, testing dact1/2 and convergent extension or wnt signaling in the ANC itself was not an aim of this paper.  

      Comment (4) The authors also have not ruled out other possibilities that could cause the dact1/2-/- ANC phenotype. For example, increased cell death or reduced proliferation in the ANC may result in the phenotype, and changes in cell fate specification or differentiation in the second CNCC stream may also result in their inability to contribute to the ANC. 

      We agree that we cannot rule out whether cell death or proliferation is different in the dact1/2 mutant ANC. However, because we do not find the second CNCC stream within the ANC, this is the most likely explanation for the abnormal ANC shape. Because the first stream of CNCC are able to populate the ANC and differentiate normally, it is most likely that the inability of the second stream to populate the ANC is due to steric hindrance imposed by the abnormal cranial/eye field morphology. These hypotheses would need to be tested, ideally with an inducible dact1/2 mutant, however, this is beyond the scope of this paper.     

      Comment (5) The last paragraph of the section "Genetic interaction of dact1/2 with Wnt regulators..." misuses terms and conflates phenotypes observed. For instance, the authors wrote "dact2 haploinsuffciency in the context of dact1-/-; gpc4-/- double mutant produced ANC in the opposite phenotypic spectrum of ANC morphology, appearing similar to the gpc4-/- mutant phenotype". However, if heterozygous dact2 is not modulating phenotypes in this genetic background, its function is not "haploinsuffcient". The authors then said, "These results show that dact1 and dact2 do not have redundant function during craniofacial morphogenesis, and that dact2 function is more indispensable than dact1". However this statement should be confined to the context of modulating gpc4 phenotypes, which is not clearly stated. 

      Revised (Lines 380, 382).   

      Comment (6) For the scRNA-seq analysis, the authors should show the population distribution in the UMAP for the 3 genotypes, even if there are no obvious changes. The authors are encouraged, although not required, to perform pseudotime or RNA velocity analysis to determine if differentiation trajectories are changed in the NC populations, in light of what they found in Fig. 4. The authors can also check the expression of reporter genes downstream of certain pathways, e.g. axin2 in canonical wnt signaling, to query if these signaling activities are changed (also related to point #3 above). 

      We have added population distribution data for the 3 genotypes to Supplemental Figure 4. Although RNA velocity analysis would be an interesting additional analysis, we would hypothesize that the NC population is not driving the differences in phenotype. Rather these are likely changes in the anterior neural plate and mesoderm. 

      Comment (7) While the phenotypic difference between gpc4-/- and dact1/2-/- are in the ANC at a later stage, ssRNA-seq was performed using younger embryos. The authors should better explain the rationale and discuss how transcriptomic differences in these younger embryos can explain later phenotypes. Importantly, dact1, dact2, and capn8 expression were not shown in and around the ANC during its development and this information is crucial for interpreting some of the results shown in this paper. For example, if dact1 and dact2 are expressed during ANC development, they may have specific functions during that stage. Alternatively, if dact1 and dact2 are not expressed when the second stream CNCCs are found to be outside the ANC, then the ANC phenotype may be due to dact1/2's functions at an earlier time point. The author's statement in the discussion that "embryonic fields determined during gastrulation effect the CNCC ability to contribute to the craniofacial skeleton" is currently speculative. 

      We have reworded our rationale and hypothesis to increase clarity (Lines 391-405). We believe that the ANC phenotype of the dact1/2 mutants is secondary to defective CE and anterior axis lengthening, as has been reported for the slb mutant (Heisenberg 1997, 2000). We utilized the gpc4 mutant as a foil to the dact1/2 mutant, as the gpc4 mutant has defective CE and axis extension without the same craniofacial phenotype.

      We have added dact1 and dact2 WISH of 24 and 48 hpf (Fig1. D,E) to show expression during ANC development. 

      Comment (8) The functional testing of capn8 did not yield a result that would suggest a strong effect, as only 1 in 142 animals phenocopied dact1/2. Therefore, while the result is interesting, the authors should tone down its importance. Alternatively, the authors can try knocking down capn8 in the dact1/2 mutants to test how that affects the CE phenotype during axis elongation, as well as ANC morphogenesis. 

      As overexpression of capn8 in wildtype animals did not result in a significant phenotype, we tested capn8 overexpression in compound dact1/2 mutants as these have a sensitized background. We found a small but statistically significant effect of exogenous capn8 in dact1+/-,dact2+/- animals. While the effect is not what one would expect comparing to Mendelian genetic ratios, the rod-like ANC phenotype is an extreme craniofacial dysmorphology not observed in wildtype or mRNA injected embryos hence significant. The experiment is limited by the available technology of over-expressing mRNA broadly without temporal or cell specificity control. It is possible that if capn8 over-expression was restricted to specific cells (floor plate, notochord or mesoderm) and at the optimal time period during gastrulation/segmentation that the aberrant ANC phenotype would be more robust. We agree with the reviewer that although the finding of a new role for capn8 during development is interesting, its importance in the context of dact should be toned down and we have altered the manuscript accordingly (Lines 455-467).  

      Comment (9) A difference between the two images in Fig. 8B is hard to distinguish.

      Consider showing flat-mount images. 

      We have added flat-mount images to Fig. 8B

      Minor comments:

      Comment (1) wnt11f2 is spelled incorrectly in a couple of places, e.g. "wnt11f2l" in the abstract and "wntllf2" in the discussion. 

      Revised throughout.

      Comment (2) For Fig. 1D, the white dact1 and yellow dact2 are hard to distinguish in the merged image. Consider changing one of their colors to a different one and only merge dact1 and dact2 without irf6 to better show their complementarity.  

      We agree with the reviewer that the expression patterns of dact1 and dact2 are difficult to distinguish in the merged image. We have added outlines of the cartilage elements to the images to facilitate comparisons of dact1 and dact2 expression (Fig 1F). 

      Comment (3) For Fig. 1E, please label the clusters mentioned in the text so readers can better compare expressions in these cell populations.  

      We have moved this data to supplementary figure S1 and have added labels.

      Comment (4) The citing and labelling of certain figures can be more specific. For example, Fig. S1A, B, and Fig. S1C should be used instead of just Fig. S1 (under the section titled dact1 and dact2 contribute to axis extension...". Similarly, Fig. 4 can be better labeled with alphabets and cited at the relevant places in the text.  

      We have modified the labeling of the figures according to the reviewer’s suggestion (Fig S2 (previously S1), Fig4) and have added reference to these labels in the text (Lines 202, 204, 212, 328, 334, 336). 

      Comment (5) For Fig. 2B, the (+/+,-/-) on x-axis should be (+/-,-/-).  

      Revised in Figure 2B.

      Comment (6) Several figures are incorrectly cited. Fig. 2C is not cited, and the "Fig. 2C" and "Fig. 2D" cited in the text should be "Fig. 2D" and "Fig. 2E" respectively. Similarly, Fig. 5C and D are not cited in the text and the cited Fig. 5C should be 5E. The VC images in Fig. 5 are not talked about in the text. Finally, Fig. 7C was also not mentioned in the text.  

      We have corrected the labeling and have added descriptions of each panel in the Results (Fig.2 Line 231, 237, 242, Fig 5 Line 373, 381, Fig 7 line 431). 

      Comment (7) In the main text, it is indicated that zebrafish at 3ss were used for ssRNAseq, but in the figure legend, it says 4ss. 

      Revised (Line 682)

      Comment (8) No error bars in Fig. S1B and the difference between the black and grey shades in Fig. S1D is not explained.  

      Error bars are not included in the graphs of qPCR results (now Fig S2C) as these are results of a pool of 8 embryos performed one time. We have added a legend to explain the gray vs. black bars (now Fig S2E). 

      Reviewer #3 (Public Review):  

      Weaknesses: The hypotheses are very poorly defined and misinterpret key previous findings surrounding the roles of wnt11 and gpc4, which results in a very confusing manuscript. Many of the results are not novel and focus on secondary defects. The most novel result of overexpressing calpain8 in dact1/2 mutants is preliminary and not convincing.  

      We apologize for not presenting the question more clearly. The Introduction was revised with particular attention to distinguish this work using genetic germline mutants from prior morpholino studies. Please refer to pages 4-5, lines 106-121.

      Weakness 1) One major problem throughout the paper is that the authors misrepresent the fact that wnt11f2 and gpc4 act in different cell populations at different times. Gastrulation defects in these mutants are not similar: wnt11 is required for anterior mesoderm CE during gastrulation but not during subsequent craniofacial development while gpc4 is required for posterior mesoderm CE and later craniofacial cartilage morphogenesis (LeClair et al., 2009). Overall, the non-overlapping functions of wnt11 and gpc4, both temporally and spatially, suggest that they are not part of the same pathway.  

      We have reworded the text to add clarity. While the loss of wnt11 versus the loss of gpc4 may affect different cell populations, the overall effect is a shortened body axis. We stressed that it is this similar impaired axis elongation phenotype but discrepant ANC morphology phenotypes in the opposite ends of the ANC morphologic spectrum that is very interesting and leads us to investigate dact1/2 in the genetic contexts of wnt11f2 and gpc4.  Pls refer to page 4, lines 73-84. Further, the reviewer’s comment that wnt11 and gpc4 are spatially and temporally distinct is untested. We think the reviewer’s claim of gpc4 acting in the posterior mesoderm refers to its requirement in the tailbud (Marlow 2004). However this does not exclude gpc4 from acting elsewhere as well. Further experiments would be necessary. Both wnt11f2 and gpc4 regulate non-canonical wnt signaling and are coexpressed during some points of gastrulation and CF development (Gupta et al., 2013; Sisson 2015). This data supports the possibility of overlapping roles. 

      Weakness 2) There are also serious problems surrounding attempts to relate single-cell data with the other data in the manuscript and many claims that lack validation. For example, in Fig 1 it is entirely unclear how the Daniocell scRNA-seq data have been used to compare dact1/2 with wnt11f2 or gpc4. With no labeling in panel 1E of this figure these comparisons are impossible to follow. Similarly, the comparisons between dact1/2 and gpc4 in scRNA-seq data in Fig. 6 as well as the choices of DEGs in dact1/2 or gpc4 mutants in Fig. 7 seem arbitrary and do not make a convincing case for any specific developmental hypothesis. Are dact1 and gpc4 or dact2 and wnt11 coexpressed in individual cells? Eyeballing similarity is not acceptable.  

      We have moved the previously published Daniocell data to Figure S1 and have added labeling. These data are meant to complement and support the WISH results and demonstrate the utility of using available public Daniocell data. Please recommend how we can do this better or recommend how we can remediate this work with specific comment. 

      Regarding our own scRNA-seq data, we have added rationale (line 391-403) and details of the results to increase clarity (Lines 419-436). We have added a panel to Figure 6 (panel A) to help illustrate or rationale for comparing dact1/2 to gpc4 mutants to wt. The DEGs displayed in Fig.7A are the top 50 most differentially expressed genes between dact1/2 mutants and WT (Figure 7 legend, line 422-424).   

      We have looked at our scRNA-seq gene expression results for our clusters of interest (lateral plate mesoderm, paraxial mesoderm, and ectoderm). We find dact1, dact2, and gpc4 co-expression within these clusters. Knowing whether these genes are coexpressed within the same individual cell would require going back and analyzing the raw expression data. We do not find this to be necessary to support our conclusions. The expression pattern of wnt11f2 is irrelevant here.   

      Weakness 3) Many of the results in the paper are not novel and either confirm previous findings, particularly Waxman et al (2004), or even contradict them without good evidence. The authors should make sure that dact2 loss-of-function is not compensated for by an increase in dact1 transcription or vice versa. Testing genetic interactions, including investigating the expression of wnt11f2 in dact1/2 mutants, dact1/2 expression in wnt11f2 mutants, or the ability of dact1/2 to rescue wnt11f2 loss of function would give this work a more novel, mechanistic angle.

      We clarified here that the prior work carried out by Waxman using morppholinos, while acceptable at the time in 2004, does not meet the rigor of developmental studies today which is to generate germline mutants. The reviewer’s acceptance of the prior work at face value fails to take the limitation of prior work into account. Further, the prior paper from Waxman et al did not analyze craniofacial morphology other than eyeballing the shape of the head and eyes. Please compare the Waxman paper and this work figure for figure and the additional detail of this study should be clear. Again, this is by no means any criticism of prior work as the prior study suffered from the technological limitations of 2004, just as this study also is the best we can do using the tools we have today. Any discrepancies in results are likely due to differences in morpholino versus genetic disruption and most reviewers would favor the phenotype analysis from the germline genetic context. We have addressed these concerns as objectively as we can in the text (Lines 482-493). The fact that dact1/2 double mutants display a craniofacial phenotype while the single mutants do not, suggests compensation (Lines 503-505), but not necessarily at the mRNA expression level (Fig. S2C). 

      This paper tests genetic interaction through phenotyping the wntll/dact1/dact2 mutant.

      Our results support the previous literature that dact1/2 act downstream of wnt11 signaling. There is no evidence of cross-regulation of gene expression. We do not expect that changes in wnt11 or dact would result in expression changes in the others.

      RNA-seq of the dact1/2 mutants did not show changes in wnt11 gene expression. Unless dact1 and/or dact2 mRNA are under expressed in the wnt11 mutant, we would not expect a rescue experiment to be informative. And as wnt11 is not a focus of this paper, we have not performed the experiment.  

      Weakness 4) The identification of calpain 8 overexpression in Dact1/2 mutants is interesting, but getting 1/142 phenotypes from mRNA injections does not meet reproducibility standards.

      As the occurrence of the mutant phenotype in wildtype animals with exogenous capn8 expression was below what would meet reproducibility standards, we performed an additional experiment where capn8 was overexpressed in embryos resulting from dact1/dact2 double heterozygotes incross (Fig. 8). We reasoned that an effect of capn8 overexpression may be more robust on a sensitized background. We found a statistically significant effect of capn8 in dact1/2 double heterozygotes, though the occurrence was still relatively rare (6/80). These data suggest dysregulation of capn8 contributes to the mutant ANC phenotype, though there are likely other factors involved. 

      Comment: The manuscript title is not representative of the findings of this study.  

      We revised the title to strictly describe that we generated and carried out genetic analysis in loss of function compound mutants (Genetic requirement) and that we found capn8 was important which modified this requirement.

      Introduction: p.4:

      Comment: Anterior neurocranium (ANC) - it has to be stated that this refers to the combined ethmoid plate and trabecular cartilages. 

      Thank you, we agree that the ANC and ethmoid plate terminology has been confusing in the literature and we should endeavor to more clearly describe that the phenotypes in question are all in the ethmoid plate and the trabeculae are not affected. ANC has been replaced with ethmoid plate (EP) throughout the manuscript and figures. We also describe that all the observed phenotypes affect the ethmoid plate and not the trabeculae, (pages 13, Lines 265-267).

      Comment: Transverse dimension is incorrect terminology - replace with medio-lateral.

      Revised (Lines 69, 74).

      Comment: Improper way of explaining the relationship between mutant and gene..."Another mutant knypek, later identified as gpc4..." a better  way to explain this would be that the knypek mutation was found to be a non-sense mutation in the gpc4 gene.  

      Revised (Line 71)

      Comment: "...the gpc4 mutant formed an ANC that is wider in the transverse dimension than the wildtype, in the opposite end of the ANC phenotypic spectrum compared to wnt11f2...These observations beg the question how defects in early patterning and convergent extension of the embryo may be associated with later craniofacial morphogenesis."

      This statement is broadly representative of the general failure to distinguish primary from secondary defects in this manuscript. Focusing on secondary defects may be useful to understand the etiology of a human disease, but it is misleading to focus on secondary defects when studying gene function. The rod-like ethmoid of slb mutant results from a CE defect of anterior mesoderm during gastrulation(Heisenberg et al. 1997, 2000), while the wide ethmoid plate of kny mutants results from CE defects of cartilage precursors (Rochard et al., 2016). Based on this evidence, wnt11f2 and gpc4 act in different cell populations at different times.  

      It is true that the slb mutant craniofacial phenotype has been stated as secondary to the CE defect during gastrulation and the kny phenotype as primary to chondrocyte CE defects in the ethmoid, however the direct experimental evidence to conclude only primary or only secondary effects does not yet exist. There is no experiment to our knowledge where wnt11f2 was found to not affect ethmoid chondrocytes directly. Likewise, there is no experiment having demonstrated that dysregulated CE in gpc4 mutants does not contribute to a secondary abnormality in the ethmoid. 

      Here, we are analyzing the CE and craniofacial phenotypes of the dact1/2 mutants without any assumptions about primary or secondary effects and without drawing any conclusions about wnt11f2 or gpc4 cellular mechanisms.     

      Comment: "The observation that wnt11f2 and gpc4 mutants share similar gastrulation and axis extension phenotypes but contrasting ANC morphologies supports a hypothesis that convergent extension mechanisms regulated by these Wnt pathway genes are specific to the temporal and spatial context during embryogenesis."

      This sentence is quite vague and potentially misleading. The gastrulation defects of these 2 mutants are not similar - wnt11 is required for anterior mesoderm CE during gastrulation and has not been shown to be active during subsequent craniofacial development while gpc4 is required for posterior mesoderm CE and craniofacial cartilage morphogenesis (LeClair et al., 2009). Here again, the non-spatially overlapping functions of wnt11 and gpc4 suggest that are not part of the same pathway.  

      Though the cells displaying defective CE in wnt11f2 and gpc4 mutants are different, the effects on the body axis are similar. The dact1/2 showed a similar axis extension defect (grossly) to these mutants. Our aim with the scRNA-seq experiment was to determine which cells and gene programs are disrupted in dact1/2 mutants. We found that some cell types and programs were disrupted similarly in dact1/2 mutants and gpc4 mutants, while other cells and programs were specific to dact1/2 versus gpc4 mutants. We can speculate that these that were specific to dact1/2 versus gpc4 may be attributed to CE in the anterior mesoderm, as is the case for wnt11. 

      p.5

      Comment: "We examined the connection between convergent extension governing gastrulation, body axis segmentation, and craniofacial morphogenesis." A statement focused on the mechanistic findings of this paper would be welcome here, instead of a claim for a "connection" that is vague and hard to find in the manuscript.  

      We have rewritten this statement (Line 125).

      p.7 Results:

      Comment: It is unclear why Farrel et al., 2018 and Lange et al., 2023 are appropriate references for WISH. Please justify or edit.  

      This was a mistake and has been edited (Page 9).

      Comment: " Further, dact gene expression was distinct from wnt11f2." This statement is inaccurate in light of the data shown in Fig1A and the following statements - please edit to reflect the partially overlapping expression patterns.  

      We have edited to clarify (Lines 142-143).

      p.8

      Comment: "...we examined dact1 and 2 expression in the developing orofacial tissues. We found that at 72hpf..." - expression at 72hpf is not relevant to craniofacial morphogenesis, which takes place between 48h-60hpf (Kimmel et al., 1998; Rochard et al., 2016; Le Pabic et al., 2014).  

      We have included images and discussion of dact1 and dact2 expression at earlier time points that are important to craniofacial development (Lines 160-171)(Fig 1D,E). 

      Comment: "This is in line with our prior finding of decreased dact2 expression in irf6 null embryos". - This statement is too vague. How are th.e two observations "in line".  

      We have removed this statement from the manuscript.

      Comment: Incomplete sentence (no verb) - "The differences in expression pattern between dact1 and dact2...".  

      Revised (Line 172).

      Comment: "During embryogenesis..." - Please label the named structures in Fig.1E.

      Please be more precise with the described expression time. Also, it would be useful to integrate the scRNAseq data with the WISH data to create an overall picture instead of treating each dataset separately.  

      We have moved the previously published Daniocell data to supplementary figure S1 and have labeled the key cell types. 

      p.9

      Comment: "The specificity of the gene disruption was demonstrated by phenotypic rescue with the injection of dact1 or dact2 mRNA (Fig. S1)." - please describe what is considered a phenotypic rescue.

      -The body axis reduction of dact mutants needs to be documented in a figure. Head pictures are not sufficient. Is the head alone affected, or both the head and trunk/tail? Fig.2E suggests that both head and trunk/tail are affected - please include a live embryos picture at a later stage.  

      We have added a description of how phenotypic rescue was determined (Line 208). We have added a figure with representative images of the whole body of dact1/2 mutants. Measurements of body length found a shortening in dact1/2 double mutants versus wildtype, however differences were not found to be significantly different by ANOVA (Fig. 3C, Fig. S3, Line 270-275).

      p. 11

      Comment: "These dact1-/-;dact2-/- CE phenotypes were similar to findings in other Wnt mutants, such as slb and kny (Heisenberg, Tada et al., 2000; Topczewski, Sepich et al., 2001)." The similarity between slb and kny phenotypes should be mentioned with caution as CE defects affect different regions in these 2 mutants. It is misleading to combine them into one phenotype category as wnt11 and gpc4 are most likely not acting in the same pathway based on these spatially distinct phenotypes.  

      Here we are referring to the grossly similar axis extension defects in slb and kny mutants. We refer to these mutants to illustrate that dact1 and or 2 deficiency could affect axis extension through diverse mechanisms. We have added text for clarity (Lines 249-252).  

      Comment: "No craniofacial phenotype was observed in dact1 or dact2 single mutants. However, in-crossing to generate [...] compound homozygotes resulted in dramatic craniofacial deformity."

      This result is intriguing in light of (1) the similar craniofacial phenotype previously reported by Waxman et al (2004) using morpholino- based knock-down of dact2, and the phenomenon of genetic compensation demonstrated by Jakutis and Stainier 2001 (https://doi.org/10.1146/annurev-genet-071719-020342). The authors should make sure that dact2 loss-of-function is not compensated for by an increase in dact1 transcription, as such compensation could lead to inaccurate conclusions if ignored.  

      We agree with the reviewer that genetic compensation of dact2 by dact1 likely explains the different result found in the dact2 morphant versus CRISPR mutant. We found increased dact1 mRNA expression in the dact2-/- mutant (Fig S2X) however a more thorough examination is required to draw a conclusion. Interestingly, we found that in wildtype embryos dact1 and dact2 expression patterns are distinct though with some overlap. It would be informative to investigate whether the dact1 expression pattern changes in dact2-/- mutants to account for dact2 loss.   

      Comment: "Lineage tracing of NCC movements in dact1/2 mutants reveals ANC composition" - the title is misleading - ANC composition was previously investigated by lineage tracing (Eberhardt et al., 2006; Wada et al., 2005).  

      This has been reworded (Line 292)

      p.13

      Comment: There is no frontonasal prominence in zebrafish.  

      This is true, texts have been changed to frontal prominence.  (Lines 293,

      299, 320)

      Comment: The rationale for investigating NC migration in mutants where there is a gastrula-stage failure of head mesoderm convergent extension is unclear. The whole head is deformed even before neural crest cells migrate as the eye field does not get split in two (Heisenberg et al., 1997; 2000), suggesting that the rod-like ethmoid plate is a secondary defect of this gastrula-stage defect. In addition, neural crest migration and cartilage morphogenesis are different processes, with clear temporal and spatial distinctions.  

      We carried out the lineage tracing experiment to determine which NC streams contributed to the aberrantly shaped EP, whether the anteromost NC stream frontal prominence, the second NC stream of maxillary prominence, or both.  We found that the anteromost NCC did contribute to the rod-like EP, which is different from when hedgehod signaling is disrupted,  So while it is possible that the gastrula-effect head mesoderm CE caused a secondary effect on NC migration, how the anterior NC stream and second NC stream are affected differently between dact1/2 and shh pathway is interesting.  We added discussion of this observation to the manuscript (page 23, Lines 514-520). 

      p. 14-16

      Comment: Based on the heavy suspicion that the rod-like ethmoid plate of the dact1/2 mutant results from a gastrulation defect, not a primary defect in later craniofacial morphogenesis, the prospect of crossing dact1/2 mutants with other wnt-pathway mutants for which craniofacial defects result from craniofacial morphogenetic defects is at the very least unlikely to generate any useful mechanistic information, and at most very likely to generate lots of confusion. Both predictions seem to take form here.  

      However, the ethmoid plate phenotype observed in the gpc4-/-; dact1+/-; dact2-/- mutants (Fig. 5E) does suggest that gpc4 may interact with dact1/2 during gastrulation, but that is the case only if dact1+/-; dact2-/- mutants do not have an ethmoid cartilage defect, which I could not find in the manuscript. Please clarify.  

      The perspective that the rod-like EP of the dact1/2 is due to gastrulation defect is being examined here. Why would other mutants such as wnt11f2 and gpc4 that have gastrulation CE defects have very different EP morphology, whether primary or secondary NCC effect?  Further dact1 and dact2 were reported as modifiers of Wnt signaling, so it is logical to genetically test the relationship between dact1, dact2, wnt11f2, gpc4 and wls. The experiment had to be done to investigate how these genetic combinations impact EP morphology. This study found that combined loss of dact1, dact2 and wls or gpc4 yielded new EP morphology different than those previously observed in either dact1/2, wls, gpc4, or any other mutant is important, suggesting that there are distinct roles for each of these genes contributing to facial morphology, that is not explained by CE defect alone.   

      Comment: I encourage the authors to explore ways to test whether the rod-like ethmoid of dact1/2 mutants is more than a secondary effect of the CE failure of the head mesoderm during gastrulation. Without this evidence, the phenotypes of dact1/2 -gpc4 or - wls are not going to convince us that these factors actually interact.  

      Actually, we find our results to support the hypothesis that the ethmoid of the dact1/2 mutants is a secondary effect of defective gastrulation and anterior extension of the body axis. However, our findings suggest (by contrasting to another mutant with impaired CE during gastrulation) that this CE defect alone cannot explain the dysmorphic ethmoid plate. Our single-cell RNA seq results and the discovery of dysregulated capn8 expression and proteolytic processes presents new wnt-regulated mechanisms for axis extension.    

      p. 20 Discussion

      Comment: "Here we show that dact1 and dact2 are required for axis extension during gastrulation and show a new example of CE defects during gastrulation associated with craniofacial defects."

      Waxman et al. (2004) previously showed that dact2 is involved in CE during gastrulation.

      Heisenberg et al. (1997, 2000), previously showed with the slb mutant how a CE defect during gastrulation causes a craniofacial defect.  

      The Waxman paper using morpholino to disrupt dact2 is produced limited analysis of CE and no analysis of craniofacial morphogenesis. We generated genetic mutants here to validate the earlier morpholino results and to analyze the craniofacial phenotype in detail. We have removed the word “new” to make the statement more clear (Line 475).

      Comment: "Our data supports the hypothesis that CE gastrulation defects are not causal to the craniofacial defect of medially displaced eyes and midfacial hypoplasia and that an additional morphological process is disrupted."

      It is unclear to me how the authors reached this conclusion. I find the view that medially displaced eyes and midfacial hypoplasia are secondary to the CE gastrulation defects unchallenged by the data presented. 

      This statement was removed and the discussion was reworded.

      Comment: The discussion should include a detailed comparison of this study's findings with those of zebrafish morpholino studies.  

      We have added more discussion to compare ours to the previous morpholino findings (Lines 476-484).

      Comment: The discussion should try to reconcile the different expression patterns of dact1 and dact2, and the functional redundancy suggested by the absence of phenotype of single mutants. Genetic compensation should be considered (and perhaps tested).  

      The different expression patterns of dact1 and dact2 along with our finding that dact1 and dact2 genetic deficiency differently affect the gpc4 mutant phenotype suggest that dact1 and dact2 are not functionally redundant during normal development. This is in line with the previously published data showing different phenotypes of dact1 or dact2 knockdown. However, our results that genetic ablation of both dact1 and dact2 are required for a mutant phenotype suggests that these genes can compensate upon loss of the other. This would suggest then that the expression pattern of dact1 would be changed in the dact2 mutant and visa versa. We find that this line of investigation would be interesting in future studies. We have addressed this in the Discussion (Lines 485498).

      Comment: "Based on the data...Conversely, we propose...ascribed to wnt11f2 "

      Functional data always prevail overexpression data for inferring functional requirements.  

      This is true.

      p.21

      Comment: "Our results underscore the crucial roles of dact1 and dact2 in embryonic development, specifically in the connection between CE during gastrulation and ultimate craniofacial development."

      How is this novel in light of previous studies, especially by Waxman et al. (2004) and Heisenberg et al. (1997, 2000). In this study, the authors fail to present compelling evidence that craniofacial defects are not secondary to the early gastrulation defects resulting from dact1/2 mutations.  p. 22

      We have not claimed that the craniofacial defects are not secondary to the gastrulation defects. In fact, we state that there is a “connection”. Further, we do not claim that this is the first or only such finding. We believe our findings have validated the previous dact morpholino experiments and have contributed to the body of literature concerning wnt signaling during embryogenesis. 

      Comment: The section on Smad1 discusses a result not reported in the results section. Any data discussed in the discussion section needs to be reported first in the results section.  

      We have added a comment on the differential expression of smad1 to the results section (Lines 446-448).

    1. To introduce data science, it makes sense that we ought to talk about data first. The word data is the plural of the the Latin word datum. One quick word before we continue: Because the word data is the plural of datum, I (and many people) prefer data as a plural noun—hence “What are Data?” for the section title. (In fact, I think it’s funny to define data science as “the science of datums,” but that’s a terrible joke and I promise I won’t do it again in this book). However, it’s quite common in American English to treat data as a singular word—so common in fact, that you might notice me trip up and write “What is Data?” at some point. My opinion here is strong enough that I won’t mind if you point out when I’m inconsistent but not so strong that I’m going to get picky about how you treat the word—go with whatever comes more naturally to you. Even though we rarely use the singular datum, it’s worth briefly exploring its etymology. The word means “a given”—that is, something taken for granted. That’s important: The word data was introduced in the mid-seventeenth century to supplement existing terms such as evidence and fact. Identifying information as data, rather than as either of those other two terms, served a rhetorical purpose (Poovey, 1998; Posner & Klein, 2017; Rosenberg, 2013). It converted otherwise debatable information into the solid basis for subsequent claims. Modern usage of the word data started in the 1940s and 1950s as practical electronic computers began to input, process, and output data. When computers work with data, all of that data has to be broken down to individual bits as the “atoms” that make up data. A bit is a binary unit of data, meaning that it is only capable of representing one of two values: 0 and 1. That doesn’t carry a lot of information by itself (at best, “yes” vs. “no” or TRUE vs. FALSE). However, by combining bits, we can increase the amount of information that we transmit. For example, even a combination of just two bits can express four different values: 00, 01, 10 and 11. Every time you add a new bit you double the number of possible messages you can send. So three bits would give eight options and four bits would give 16 options. When we get up to eight bits—which provides 256 different combinations—we finally have something of a reasonably useful size to work with. Eight bits is commonly referred to as a byte—this term probably started out as a play on words with the word bit (and four bits is sometimes referred to as a nibble or a nybble, because nerds like jokes). A byte offers enough different combinations to encode all of the letters of the (English) alphabet, including capital and small letters. There is an old rulebook called ASCII—the American Standard Code for Information Interchange—which matches up patterns of eight bits with the letters of the alphabet, punctuation, and a few other odds and ends. For example the bit pattern 0100 0001 represents the capital letter A and the next higher pattern 0100 0010 represents capital B. This is more background than anything else—most of the time (but not all of the time!) you don’t need to know the details of what’s going on here to carry out data science. However, it is important to have a foundational understanding that when we’re working with data in this class, the computer is ultimately dealing with everything as bits and translating combinations of bits into words, pictures, numbers, and other formats that makes sense for humans. This background is also helpful for pointing out that just like the word data has connotations related to trustworthiness, it also has connotations of things that are digital and quantitative. While all of these connotations are reasonable, it’s important that we understand their limits. For example, while many people think of data as numbers alone, data can also consist of words or stories, colors or sounds, or any type of information that is systematically collected, organized, and analyzed. Some folks might resist that broad definition of data because “words or stories” told by a person don’t feel as trustworthy or objective as numbers stored in a computer. However, one of the recurring themes of this course is to emphasize that data and data systems are not objective—even when they’re digital and quantitative. When I was introducing ASCII a few paragraphs ago, there were two details in there that might have passed you by but that actually have pretty important consequences. First, I noted that ASCII can “encode all the letters of the (English) alphabet”; second, I mentioned that the “A” in ASCII stood for “American.” Early computer systems in the United States were built around American English assumptions for what counts as a letter. This makes sense… but it has had consequences! While most modern computer systems have moved on to more advanced character encoding systems (ones that include Latin letters, Chinese characters, Arabic script, and emoji, for example), there are still some really important computer systems that use limited encoding schemes like ASCII. In 2015, Tovin Lapin wrote a newspaper article about this, noting that: Every year in California thousands of parents choose names such as José, André, and Sofía for their children, often honoring the memory of a deceased grandmother, aunt or sibling. On the state-issued birth certificates, though, those names will be spelled incorrectly. California, like several other states, prohibits the use of diacritical marks or accents on official documents. That means no tilde (~), no accent grave (`), no umlaut (¨) and certainly no cedilla (¸). Although more than a third of the state population is Hispanic, and accents are used in the names of state parks and landmarks, the state bars their use on birth records. There were attempts in 2014 to change this, but when lawmakers realized it would cost $10 million to update computer systems, things stalled. Moral of the story: even though ASCII is a straightforward technical system built on digital data with no real wiggle room for what means what, it’s still subjective and biased. How we organize data and data systems matters! So, even digital and quantitative data (systems) can be biased, which means that we ought to push lightly back against the rhetorical connotations of data as trustworthy. I’m not suggesting we throw data, science, and data science out the window and go with our gut and our opinions, but we shouldn’t take for granted that a given dataset doesn’t have its own subjectivity. Likewise, we ought to ask ourselves what information needs to become data before it can be trusted—or, more precisely, whose information needs to become data before it can be considered as fact and acted upon (Lanius, 2015; Porter, 1996).

      This section to me is the framework of what data is that will launch into doing cool things this semester definitely things to revisit for sure to get a better understanding if we don’t read it all the way through the first time

    1. Welcome back.

      In this lesson, I want to talk about S3 replication, the feature which allows you to configure the replication of objects between a source and destination S3 bucket.

      Now there are two types of replication supported by S3.

      The first type, which has been available for some time, is cross-region replication or CRR, and that allows the replication of objects from source buckets to one or more destination buckets in different AWS regions.

      The second type of replication announced more recently is same-region replication or SRR, which as the name suggests is the same process, but where both the source and destination buckets are in the same AWS region.

      Now the architecture for both types of replication is pretty simple to understand once you've seen it visually.

      It only differs depending on whether the buckets are in the same AWS accounts or different AWS accounts.

      Both types of replication support both, so the buckets could be in the same or different AWS accounts.

      In both cases, replication configuration is applied to the source bucket.

      The replication configuration configures S3 to replicate from the source bucket to a destination bucket, and it specifies a few important things.

      The first is logically the destination bucket to use as part of that replication, and another thing that's configured in the replication configuration is an IAM role to use for the replication process.

      The role is configured to allow the S3 service to assume it, so that's defined in its trust policy.

      The role's permissions policy gives it the permission to read objects on the source bucket and permissions to replicate those objects to the destination bucket.

      And this is how replication is configured between source and destination buckets, and of course that replication is encrypted.

      Now the configuration does define a few other items, but I'll talk about them on the next screen for now, and just focusing on this basic architecture.

      There is one crucial difference between replication which occurs in the same AWS accounts versus different AWS accounts.

      Inside one account, both S3 buckets are owned by the same AWS account, so they both trust that same AWS account that they're in.

      That means that they both trust IAM as a service, which means that they both trust the IAM role.

      For the same account, that means that the IAM role automatically has access to the source and the destination buckets as long as the role's permission policy grants the access.

      If you're configuring replication between different AWS accounts, though, that's nothing off.

      The destination bucket, because it's in a different AWS account, doesn't trust the source account or the role that's used to replicate the bucket contents.

      So in different accounts, remember that the role that's configured to perform the replication isn't by default trusted by the destination account because it's a separate AWS account.

      So if you're configuring this replication between different accounts, there's also a requirement to add a bucket policy on the destination bucket, which allows the role in the source account to replicate objects into it.

      So you're using a bucket policy, which is a resource policy, to define the role in a separate account can write or replicate objects into that bucket.

      Once this configuration is applied, so either the top configuration is the same account or the bottom, if it's different accounts, then S3 can perform the replication.

      Now let's quickly review some of the options available for replication configuration, so that might actually come in handy for you to know.

      The first important option is what to replicate.

      The default is to replicate an entire source bucket to a destination bucket, so all objects, all prefixes, and all packs.

      You can, though, choose a subset of objects, so you can create a rule that has a filter, and the filter can filter objects by prefix or tags or a combination of both, and that can define exactly what objects are replicated from the source to the destination.

      You can also select which storage class the objects in the destination bucket will use.

      Now the default is to use the same class, but you can pick a cheaper class if this is going to be a secondary copy of data.

      Remember when I talked about the storage classes that are available in S3, I talked about infrequent access or one-zone infrequent access classes, which could be used for secondary data.

      So with secondary data, you're able to tolerate a lower level of resiliency, so we could use one-zone infrequent access for the destination bucket objects, and we can do that because we've always got this primary copy in the source bucket, so we can achieve better economies by using a lower cost storage class in the destination.

      So remember this is the example, the default is to use the same storage class on the destination as is used on the source, but you can override that in the replication configuration.

      Now you can also define the ownership of the objects in the destination bucket.

      The default is that they will be owned by the same account as the source bucket.

      Now this is fine if both buckets are inside the same account.

      That will mean that objects in the destination bucket will be owned by the same as the source bucket, which is the same account, so that's all good.

      However, if the buckets are in different accounts, then by default the objects inside the destination bucket will be owned by the source bucket account, and that could mean when you end up in a situation where the destination account can't read those objects because they're owned by a different AWS account.

      So with this option you can override that and you can set it so that anything created in the destination bucket is owned by the destination account.

      And lastly there's an extra feature that can be enabled called replication time control or RTC, and this is a feature which adds a guaranteed 15-minute replication SLA onto this process.

      Now without this, it's a best efforts process, but RTC adds this SLA, it's a guaranteed level of predictability, and it even adds additional monitoring so you can see which objects are queued for replication.

      So this is something that you would tend to only use if you've got a really strict set of requirements from your business, and make sure that the destination bucket and source buckets are in sync as closely as possible.

      If you don't require this, if this is just performing backups or it's just for a personal project, or if the source and destination buckets aren't required to always be in sync within this 15-minute window, then it's probably not worth adding this feature.

      It's something to keep in mind and be aware of for the exam.

      If you do see any questions that mention 15 minutes for replication, then you know that you need this replication time control.

      Now there are some considerations that you should be aware of, especially for the exam.

      These will come up in the exam, so please pay attention and try to remember these points.

      The first thing is that by default replication isn't retroactive.

      You enable replication on a pair of buckets of source and destination, and only from that point onward are objects replicated from source to destination.

      So if you enable replication on a bucket which already has objects, those objects will not be replicated.

      And related to this, in order to enable replication on a bucket, both the source and destination bucket need to have versioning enabled.

      You'll be allowed to enable versioning as part of the process of enabling replication, but it is a requirement to have it on, so a bucket cannot be enabled for replication without versioning.

      Now you can use S3 batch replication to replicate existing objects, but this is something that you need to specifically configure.

      If you don't, just remember by default replication is not retroactive.

      Secondly, it is a one-way replication process only.

      Objects are replicated from the source to the destination.

      If you add objects manually in the destination, they will not be replicated back to the source.

      This is not a bi-directional replication.

      It's one-way only.

      Now more recently, AWS did add the feature which allows you to add bi-directional replication, but just be aware that this is an additional setting which you need to configure.

      By default, replication is one-way, and enabling bi-directional replication is something you need to specifically configure, so keep that in mind.

      Now in terms of what does get replicated from source to destination, replication is capable of handling objects which are unencrypted, so if you don't have any encryption on an object, it's capable of handling objects which are encrypted using S3 and S3, and it's even capable of handling objects which are encrypted using S3 and KMS, but this is an extra piece of configuration that you'll need to enable.

      So there's configuration and there's extra permissions which are required, because of course KMS is emolved.

      Now more recently, AWS did add the ability to replicate objects encrypted with SSE-C, so that's server-side encryption with custom-managed keys, but this is a relatively recent addition.

      Historically, SSE-C was incompatible with cross or same-region replication.

      Replication also requires that the owner of the source bucket needs permissions on the objects which will replicate, so in most cases, if you create a bucket in an account and you add those objects, then the owner of the object will be the source account, but if you grant cross-account access to a bucket, if you add a resource policy allowing all the AWS accounts to create objects in a bucket, it's possible that the source bucket account will not own some of those objects, and the style of replication can only replicate objects where the source account owns those objects.

      So keep that in mind, and the limitation is it will not replicate system events, so if any changes are made in the source bucket by lifecycle management, they will not be replicated to the destination bucket, so only user events are replicated, and in addition to that, it can't replicate any objects inside a bucket that are using the Belysia or Glacia Deep Archive storage classes.

      Now that makes sense, because Glacia and Belysia Deep Archive, while they are shown as being inside an S3 bucket, you need to conceptually think of them as separate storage products, so they cannot be replicated using this process.

      And then lastly, it's important to understand that by default, deletes are not replicated between buckets, so the adding of a delete marker, which is how object deletions are handled for a version-enabled bucket, by default, these delete markers are not replicated.

      Now you can enable that, but you need to be aware that by default, this isn't enabled.

      So one of the important things I need to make sure you're aware of in terms of replication is why you would use replication.

      What are some of the scenarios that you'll use replication for?

      So for the same region replication specifically, you might use this process for log aggregation, so if you've got multiple different S3 buckets which store logs for different systems, then you could use this to aggregate those logs into a single S3 bucket.

      You might want to use the same region replication to configure some sort of synchronization between production and test accounts.

      Maybe you want to replicate data from prod to test periodically, or maybe you want to replicate some testing data into your prod account.

      This can be configured in either direction, but a very common use case for same region replication is this replication between different AWS accounts, different functions, so prod and test, or different functional teams within your business.

      You might want to use same region replication to implement resilience if you have strict sovereignty requirements.

      So there are companies in certain sectors which cannot have data leaving a specific AWS region because of sovereignty requirements, so you can have same region replication replicating between different buckets and different accounts, and then you have this account isolation for that data.

      So having a separate account with separate logins isolated to make an audit team or a security team replicates it into that account, it provides this account level isolation.

      Obviously, if you don't have those sovereignty requirements, then you can use cross region replication and use replication to implement global resilience improvements, so you can have backups of your data copied to different AWS regions to cope with large scale failure.

      You can also replicate data into different regions to reduce latency.

      So if you have, for example, a web application or your application loads data, then obviously it might be latency sensitive, so you can replicate data from one AWS region to another, so the customers in that remote region can access the bucket that's closest to them, and that reduces latency generally gives them better performance.

      Now that is everything I want to cover in this video, so go ahead and complete the video, and when you're ready, I look forward to you joining me in the next.

    1. "A few weeks ago, we hosted a little dinner in New York, and we just asked this question of 20-plus CDOs [chief data officers] in New York City of the biggest companies, 'Hey, is this an issue?' And the resounding response was, 'Yeah, it's a real mess.'" Asked how many had grounded a Copilot implementation, Berkowitz said it was about half of them. Companies, he said, were turning off Copilot software or severely restricting its use. "Now, it's not an unsolvable problem," he added. "But you've got to have clean data and you've got to have clean security in order to get these systems to really work the way you anticipate. It's more than just flipping the switch."

      Companies, half of an anecdotal sample of some 20 US CDOs, have turned Copilot off / restricting it strongly. This as it surfaces info in summaries etc that employees would not have direct access to. No access security connection between Copilot and results. So data governance is blocking its roll-out.

  2. Aug 2024
    1. I think it's it's critical for us uh when for for for for people to realize that when we reimagine what the self is and take away take take us away from this this notion of a of a subst you know some kind of monatic substance and all that um it's different than what you said before which is uh that well it's you know every everything is equally illusory I mean there's there's nothing at that point well if it's that that's a deeply destabilizing concept for a lot of people

      for - question - what would Federic Faggin think of this? - question - multi-scale communication - question - are Tibetan Rainbow body and knowing time of death examples of multi-scale communications? question - what would Federic Faggin think of this? - He comes from an experiential perspective, not just an intellectual one.

      question - what would Federic Faggin think of this? - I don't think Michael Levin provides a satisfactory answer to this and this is related to the meaning crisis modernity finds itself in - when traditional religions no longer suffice, - but there is nothing in modernity that can fill the gap yet, if mortality salience is a big issue - I don't think an intellectual answer can meet the needs of people suffering in the meaning crisis, although it is necessary, it is not sufficient - I think they are after some kind of nonverbal, nondual transformative experience

      question - multi-scale communication - This is also a question about multi-scale communication - I've recently used a metaphor to compare - the unitary, monatic experience of consciousness to - an elected government - The trillions of cells "elect" consciousness" as the high level government to oversea them - but we seem to be in the situation of the government being out of touch with the citizens - At one time in our history, was it common to be able for - high level consciousness to communicate directly with - low level cells and subcellular structures? - If so, why has this practice disappeared and - how can we re-establish it?

      question - Are Tibetan Rainbow body and knowing time of death examples of multi-scale communications? - In some older spiritual traditions such as found in the East, it seems deep meditative practitioners are able to achieve a degree of communications with parts of their body that is unconventional and surprising to modern researchers - For example, Tibetan meditators report of having the abiity to predict the time of their death by recognizing subtle bodily, interoceptive signals - Rare instances also occur of the Rainbow Body, when great meditators in the Dzogchen tradition whose body at time of death can disappear in a body of light

    1. It’s critical that you pick an angle for your mailing list. I see so many authors and brands just put up a form that says: “Sign up for my newsletter.” What newsletter? Why would I want a random newsletter from a random person? I decided on book recommendations because no one else was doing it and it was something I was good at. What are you good at? What can only you offer via email?

      Excellent point

    1. Include EXTENSION REQUEST in your subject. If you ask before the due date I will almost certainly say yes, so just ask! If the due date has passed, the answer will be no.

      The extension policy is fair and reasonable. I agree that it's important for both the student and the professor to be organized and informed in advance. This also allows both to benefit and maintain a good relationship. Particularly, I wanted to thank the extension granted due to my flight cancellation!

    1. But if it’s the only value that defines a life, then students don’t need a true education at all. They don’t need to construct a vision of the whole world and their place in it. They don’t need to address the larger questions that arise through open-ended discussion with professors and peers. They need just narrowly focused training.

      Opposing view against the teaching values professors have installed to students.

    2. But if it’s the only value that defines a life, then students don’t need a true education at all. They don’t need to construct a vision of the whole world and their place in it. They don’t need to address the larger questions that arise through open-ended discussion with professors and peers. They need just narrowly focused training.

      -to desire a life with work and no enjoyment cannot even be considered living at all. Further they cannot live to a full potential in thinking in such a linear way

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      In this study, the authors used a stopped-flow method to investigate the kinetics of substrate translocation through the channel in hexameric ClpB, an ATP-dependent bacterial protein disaggregase. They engineered a series of polypeptides with the N-terminal RepA ClpB-targeting sequence followed by a variable number of folded titin domains. The authors detected translocation of the substrate polypeptides by observing the enhancement of fluorescence from a probe located at the substrate's C-terminus. The total time of the substrates' translocation correlated with their lengths, which allowed the authors to determine the number of residues translocated by ClpB per unit time.

      Strengths:

      This study confirms a previously proposed model of processive translocation of polypeptides through the channel in ClpB. The novelty of this work is in the clever design of a series of kinetic experiments with an engineered substrate that includes stably folded domains. This approach produced a quantitative description of the reaction rates and kinetic step sizes. Another valuable aspect is that the method can be used for other translocases from the AAA+ family to characterize their mechanism of substrate processing.

      Weaknesses:

      The main limitation of the study is in using a single non-physiological substrate of ClpB, which does not replicate the physical properties of the aggregated cellular proteins and includes a non-physiological ClpB-targeting sequence. Another limitation is in the use of ATPgammaS to stimulate the substrate processing. It is not clear how relevant the results are to the ClpB function in living cells with ATP as the source of energy, a multitude of various aggregated substrates without targeting sequences that need ClpB's assistance, and in the presence of the co-chaperones.

      Indeed, we agree that our RepA-Titinx substrates are not aggregates but are model, soluble, substrates used to reveal information about enzyme catalyzed protein unfolding and translocation.  Our substrates are similar to RepA-GFP and GFP-SsrA used by multiple labs including Wickner, Horwich, Sauer, Baker, Shorter, Bukua, to name only a few.  The fact that “this is what everyone does” does not make the substrates physiological or the most ideal. However, this is the technology we currently have until we and others develop something better. In the meantime, we contend that  the results presented here do advance our knowledge on enzyme catalyzed protein unfolding

      Part of what this manuscript seeks to accomplish is presenting the development of a single-turnover experiment that reports on processive protein unfolding by AAA+ molecular motors, in this case, ClpB.  Importantly, we are treating translocation on an unfolded polypeptide chain and protein unfolding of stably folded proteins as two distinct reactions catalyzed by ClpB. If these functions are used to disrupt protein aggregates, in vivo, then this remains to be seen.

      We contend that processive ClpB catalyzed protein unfolding has not been rigorously demonstrated prior to our results presented here.  Avellaneda et al mechanically unfolded their substrate before loading ClpB (Avellaneda, Franke, Sunderlikova et al. 2020).  Thus, their experiment represents valuable observations reflecting polypeptide translocation on a pre-unfolded protein.  Our previous work using single-turnover stopped-flow experiments employed unstructured synthetic polypeptides and therefore reflects polypeptide translocation and not protein unfolding (Li, Weaver, Lin et al. 2015).  Weibezahn et al used unstructured substrates in their study with ClpB (BAP/ClpP), and thus their results represent translocation of a pre-unfolded polypeptide and not enzyme catalyzed protein unfolding (Weibezahn, Tessarz, Schlieker et al. 2004). 

      Many studies have reported the use of  GFP with tags or RepA-GFP and used the loss of GFP fluorescence to conclude protein unfolding.  However, such results do not reveal if ClpB processively and fully translocates the substrate through its axial channel.  One cannot rule out, even when trapping with “GroEL trap”, the possibility that ClpB only needs to disrupt some of the fold in GFP before cooperative unfolding occurs leading to loss of fluorescence.  Once the cooperative collapse of the structure occurs and fluorescence is lost it has not been shown that ClpB will continue to translocate on the newly unfolded chain or dissociate. In fact, the Bukau group showed that folded YFP remained intact after luciferase was unfolded (Haslberger, Zdanowicz, Brand et al. 2008).  Our approach, reported here, yields signal upon arrival of the motor at the c-terminus or within the PIFE distance thus we can be certain that the motor does arrive at the c-terminus after unfolding up to three tandem repeats of the Titin I27 domain.

      ATPgS is a non-physiological nucleotide analog.  However, ClpB has been shown to exhibit curious behavior in its presence that we and others, as the reviewer acknowledges, do not fully understand (Doyle, Shorter, Zolkiewski et al. 2007).  Some of the experiments reported here are seeking to better understand that fact.  Here we have shown that ATPgS alone will support processive protein unfolding. With this assay in hand, we are now seeking to go forward and address many of the points raised by this reviewer. 

      The authors do not attempt to correlate the kinetic step sizes detected during substrate translocation and unfolding with the substrate's structure, which should be possible, given how extensively the stability and unfolding of the titin I27 domain were studied before. Also, since the substrate contains up to three I27 domains separated with unstructured linkers, it is not clear why all the translocation steps are assumed to occur with the same rate constant.

      We assume that all protein unfolding steps occur with the same rate constant, ku.  We conclude that we are not detecting the translocation rate constant, kt, as our results support a model where kt is much faster than ku.  We do think it makes sense that the same slow step occurs between each cycle of protein unfolding.

      We have added a discussion relating our observations to mechanical unfolding of tandem repeats of Titin I27 from AFM experiments  (Oberhauser, Hansma, Carrion-Vazquez and Fernandez 2001). Most interestingly, they report unfolding of Titin I27 in 22 nm steps.  Using 0.34 nm per amino acids this yields ~65 amino acids per unfolding step, which is comparable to our kinetic step-size of 57 – 58 amino acids per step.

      Some conclusions presented in the manuscript are speculative:

      The notion that the emission from Alexa Fluor 555 is enhanced when ClpB approaches the substrate's C-terminus needs to be supported experimentally. Also, evidence that ATPgammaS without ATP can provide sufficient energy for substrate translocation and unfolding is missing in the paper.

      In our previous work we have used fluorescently labeled 50 amino acid peptides as substrates to examine ClpB binding (Li, Lin and Lucius 2015, Li, Weaver, Lin et al. 2015).  In that work we have used fluorescein, which exhibits quenching upon ClpB binding.  We have added a control experiment where we have attached alexa fluor 555 to the 50 amino acid substrate so we can be assured the ClpB binds close to the fluorophore.  As seen in supplemental Fig. 1 A  upon titration with ClpB, in the presence of ATPγS, we observe an increase in fluorescence from AF555, consistent with PIFE.  Supplemental Fig. 1 B shows the relative fluorescence enhancement at the peak max increases up to ~ 0.2 or a 20 % increase in fluorescence, due to PIFE, upon ClpB binding.   

      Further, peak time is our hypothesized measure of ClpB’s arrival at the dye. Our results indicate that the peak time linearly increases as a function of an increase in the number of folded TitinI27 repeats in the substrates which also supports the PIFE hypothesis. Finally, others have shown that AF555 exhibits PIFE and we have added those references.

      The evidence that ATPγS alone can support translocation is shown in Fig. 2 and supplemental Figure 1.  Fig. 2 and supplemental Figure 1 are two different mixing strategies where we use only ATPgS and no ATP at all.  In both cases the time courses are consistent with processive protein unfolding by ClpB with only ATPγS.

      Reviewer #2 (Public Review):

      Summary:

      The current work by Banwait et al. reports a fluorescence-based single turnover method based on protein-induced fluorescence enhancement (PIFE) to show that ClpB is a processive motor. The paper is a crucial finding as there has been ambiguity on whether ClpB is a processive or non-processive motor. Optical tweezers-based single-molecule studies have shown that ClpB is a processive motor, whereas previous studies from the same group hypothesized it to be a non-processive motor. As co-chaperones are needed for the motor activity of the ClpB, to isolate the activity of ClpB, they have used a 1:1 ratio ATP and ATPgS, where the enzyme is active even in the absence of its co-chaperones, as previously observed. A sequential mixing stop-flow protocol was developed, and the unfolding and translocation of RepA-TitinX, X = 1,2,3 repeats was monitored by measuring the fluorescence intensity with the time of Alexa F555 which was labelled at the C-terminal Cysteine. The observations were a lag time, followed by a gradual increase in fluorescence due to PIFE, and then a decrease in fluorescence plausibly due to the dissociation from the substrate allowing it to refold. The authors observed that the peak time depends on the substrate length, indicating the processive nature of ClpB. In addition, the lag and peak times depend on the pre-incubation time with ATPgS, indicating that the enzyme translocates on the substrates even with just ATPgS without the addition of ATP, which is plausible due to the slow hydrolysis of ATPgS. From the plot of substrate length vs peak time, the authors calculated the rate of unfolding and translocation to be ~0.1 aas-1 in the presence of ~1 mM ATPgS and increases to 1 aas-1 in the presence of 1:1 ATP and ATPgS. The authors have further performed experiments at 3:1 ATP and ATPgS concentrations and observed ~5 times increase in the translocation rates as expected due to faster hydrolysis of ATP by ClpB and reconfirming that processivity is majorly ATP driven. Further, the authors model their results to multiple sequential unfolding steps, determining the rate of unfolding and the number of amino acids unfolded during each step. Overall, the study uses a novel method to reconfirm the processive nature of ClpB.

      Strengths:

      (1) Previous studies on understanding the processivity of ClpB have primarily focused on unfolded or disordered proteins; this study paves new insights into our understanding of the processing of folded proteins by ClpB. They have cleverly used RepA as a recognition sequence to understand the unfolding of titin-I27 folded domains.

      (2) The method developed can be applied to many disaggregating enzymes and has broader significance.

      (3) The data from various experiments are consistent with each other, indicating the reproducibility of the data. For example, the rate of translocation in the presence of ATPgS, ~0.1 aas-1 from the single mixing experiment and double mixing experiment are very similar.

      (4) The study convincingly shows that ClpB is a processive motor, which has long been debated, describing its activity in the presence of only ATPgS and a mixture of ATP and ATPgS.

      (5) The discussion part has been written in a way that describes many previous experiments from various groups supporting the processive nature of the enzyme and supports their current study.

      Weaknesses:

      (1) The authors model that the enzyme unfolds the protein sequentially around 60 aa each time through multiple steps and translocates rapidly. This contradicts our knowledge of protein unfolding, which is generally cooperative, particularly for titinI27, which is reported to unfold cooperatively or utmost through one intermediate during enzymatic unfolding by ClpX and ClpA.

      We do not think this represents a contradiction.  In fact, our observations are in good agreement with mechanical unfolding of tandem repeats of Titin I27 using AFM experiments (Oberhauser, Hansma, Carrion-Vazquez and Fernandez 2001).  They showed that tandem repeats of TitinI27 unfolded in steps of ~22 nm.  Dividing 22 nm by 0.34 nm/Amino Acid gives ~65 amino acids per unfolding event.  This implies that, under force, ~65 amino acids of folded structure unfolds in a single step.  This number is in excellent agreement with our kinetic step-size of 65 AA/step. 

      Importantly, the experiments cited by the reviewer on ClpA and ClpX are actually with ClpAP and ClpXP.  We assert that this is an important distinction as we have shown that ClpA employs a different mechanism than ClpAP (Rajendar and Lucius 2010, Miller, Lin, Li and Lucius 2013, Miller and Lucius 2014).  Thus, ClpA and ClpAP should be treated as different enzymes but, without question, ClpB and ClpA are different enzymes.

      (2) It is also important to note that the unfolding of titinI27 from the N-terminus (as done in this study) has been reported to be very fast and cannot be the rate-limiting step as reported earlier(Olivares et al, PNAS, 2017). This contradicts the current model where unfolding is the rate-limiting step, and the translocation is assumed to be many orders faster than unfolding.

      Most importantly, the Olivares paper is examining ClpXP and ClpAP catalyzed protein unfolding and translocation and not ClpB.  These are different enzymes.  Additionally, we have shown that ClpAP and ClpA translocate unfolded polypeptides with different rates, rate constants, and kinetic step-sizes indicating that ClpP allosterically impacts the mechanism employed by ClpA to the extent that even ClpA and ClpAP should be considered different enzymes (Rajendar and Lucius 2010, Miller, Lin, Li and Lucius 2013).  We would further assert that there is no reason to assume ClpAP and ClpXP would catalyze protein unfolding using the same mechanism as ClpB as we do not think it should be assumed ClpA and ClpX use the same mechanism as ClpAP and ClpXP, respectively. 

      The Olivares et al paper reports a dwell time preceding protein unfolding of ~0.9 and ~0.8 s for ClpXP and ClpAP, respectively.   The inverse of this can be taken as the rate constant for protein unfolding and would yield a rate constant of ~1.2 s-1, which is in good agreement with our observed rate constant of 0.9 – 4.3 s-1 depending on the ATP:ATPγS mixing ratio.  For ClpB, we propose that the slow unfolding is then followed by rapid translocation on the unfolded chain where translocation by ClpB must be much faster than for ClpAP and ClpXP.  We think this is a reasonable interpretation of our results and not a contradiction of the results in Olivares et al. Moreover, this is completely consistent with the mechanistic differences that we have reported, using the same single-turnover stopped flow approach on the same unfolded polypeptide chains with ClpB, ClpA, and ClpAP (Rajendar and Lucius 2010, Miller, Lin, Li and Lucius 2013, Miller and Lucius 2014, Li, Weaver, Lin et al. 2015).

      (3) The model assumes the same time constant for all the unfolding steps irrespective of the secondary structural interactions.

      Yes, we contend that this is a good assumption because it represents repetition of protein unfolding catalyzed by ClpB upon encountering the same repeating structural elements, i.e. Beta sheets. 

      (4) Unlike other single-molecule optical tweezer-based assays, the study cannot distinguish the unfolding and translocation events and assumes that unfolding is the rate-limiting step.

      Although we cannot, directly, distinguish between protein unfolding and translocation we have logically concluded that protein unfolding is likely rate limiting. This is because the large kinetic step-size represents the collapse of ~60 amino acids of structure between two rate-limiting steps, which we interpret to represent cooperative protein unfolding induced by ClpB.  It is not an assumption it is our current best interpretation of the observations that we are now seeking to further test. 

      Reviewer #3 (Public Review):

      Summary:

      The authors have devised an elegant stopped-flow fluorescence approach to probe the mechanism of action of the Hsp100 protein unfoldase ClpB on an unfolded substrate (RepA) coupled to 1-3 repeats of a folded titin domain. They provide useful new insight into the kinetics of ClpB action. The results support their conclusions for the model setup used.

      Strengths:

      The stopped-flow fluorescence method with a variable delay after mixing the reactants is informative, as is the use of variable numbers of folded domains to probe the unfolding steps.

      Weaknesses:

      The setup does not reflect the physiological setting for ClpB action. A mixture of ATP and ATPgammaS is used to activate ClpB without the need for its co-chaperones, Hsp70. Hsp40 and an Hsp70 nucleotide exchange factor. This nucleotide strategy was discovered by Doyle et al (2007) but the mechanism of action is not fully understood. Other authors have used different approaches. As mentioned by the authors, Weibezahn et al used a construct coupled to the ClpA protease to demonstrate translocation. Avellaneda et al used a mutant (Y503D) in the coiled-coil regulatory domain to bypass the Hsp70 system. These differences complicate comparisons of rates and step sizes with previous work. It is unclear which results, if any, reflect the in vivo action of ClpB on the disassembly of aggregates.

      We agree with the reviewer, there are several strategies that have been employed to bypass the need for Hsp70/40 or KJE to simplify in vitro experiments.  Here we have developed a first of its kind transient state kinetics approach that can be used to examine processive protein unfolding.  We now seek to go forward with examining the mechanisms of hyperactive mutants, like Y503D, and add the co-chaperones so that we can address the limitations articulated by the reviewer.   In fact we already began adding DnaK to the reaction and found that DnaK induced ClpB to release the polypeptide chain (Durie, Duran and Lucius 2018).  However, the sequential mixing strategy developed here was needed to go forward with examining the impact of co-chaperones. 

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors):

      Line 1: I recommend changing the title of the paper to remove the terms that are not clearly defined in the text: "robust" and "processive". What are the Authors' criteria for describing a molecular machine as "robust" vs. "not robust"? A definition of processivity is given in equation 2, but its value for ClpB is not reported in the text, and the criteria for classifying a machine as "processive" vs. "non-processive" are not included. Besides, the Authors have previously reported that ClpB is non-processive (Biochem. J., 2015), so it is now clear that a more nuanced terminology should be applied to this protein. Also, Escherichia coli should be fully spelled out in the title.

      The title has been changed.  We have removed “robust” as we agree with the reviewer, there is no way to quantify “robust”.  However, we have kept “processive” and have added to the discussion a calculation of processivity since we can quantify processivity.  Importantly, the unstructured substrates used in our previous studies represent translocation and not protein unfolding.  here, on folded substrates, we detect rate-limiting protein unfolding followed by rapid translocation.  Thus, we report a lower bound on protein unfolding processivity of 362 amino acids. 

      Line 20: The comment about mitochondrial SKD3 should be removed. SKD3, like ClpB, belongs to the AAA+ family, and it is simply a coincidence that the original study that discovered SKD3 termed it an Hsp100 homolog. The similarity between SKD3 and ClpB is limited to the AAA+ module, so there are many other metazoan ATPases, besides SKD3, that could be called homologs of ClpB, including mitochondrial ClpX, ER-localized torsins, p97, etc.

      Removed.

      Lines 133-139. Contrary to what the authors state, it is not clear that the "lag-phase" becomes significantly shorter for subsequent mixing experiments (Figure 1E) perhaps except for the last one (2070 s). It is clear, however, that the emission enhancement becomes stronger for later mixes. This effect should be discussed and explained, as it suggests that the pre-equilibrations shorter than ~2000 sec do not produce saturation of ClpB binding to the substrate.

      We have added supplemental figure 2, which represents a zoom into the lag region.  This better illustrates what we were seeing but did not clearly show to the reader.  In addition, we address all three changes in the time courses, i.e. extend of lag, change in peak position, and the change in peak height. 

      Line 175. The hydrolysis rate of ATPgammaS in the presence of ClpB should be measured and compared to the hydrolysis rate with ATP/ATPgammaS to check if the ratio of those rates agrees with the ratio of the translocation rates. These experiments should be performed with and without the RepA-titin substrate, which could reveal an important linkage between the ATPase engine and substrate translocation. These experiments are essential to support the claim of substrate translocation and unfolding with ATPgammaS as the sole energy source.

      The time courses shown in figure 2 and supplemental Figure 1 are collected with only ATPgS and no ATP.  The time courses show a clear increase in lag and appearance of a peak with increasing number of tandem repeats of titin domains.  We do not see an alternate explanation for this observation other than ATPγS supports ClpB catalyzed protein unfolding and translocation.  What is the reviewers alternate explanation for these observations?

      We agree with the reviewer that the linkage of ATP hydrolysis to protein unfolding and translocation is essential and we are seeking to acquire this knowledge.  However, a simple comparison of the ratio of rates is not adequate. We contend that a complete mechanistic study of ATP turnover by ClpB is required to properly address this linkage and such a study is too substantial to be included here but is currently underway. 

      All that said, the statement on line 175 was removed since we do not report any ATPase measurements in this paper.

      Line 199: It is an over-simplification to state that "1:1 mix of ATP to ATPgammaS replaces the need for co-chaperones". This sentence should be corrected or removed. The ClpB co-chaperones (DnaK, DnaJ, GrpE) play a major role in targeting ClpB to its aggregated substrates in cells and in regulating the ClpB activity through interactions with its middle domain. ATPgammaS does not replace the co-chaperones; it is a chemical probe that modifies the mechanism of ClpB in a way that is not entirely understood.

      We agree with the reviewer.  The sentence has been modified to point out that the mix of ATP and ATPγS activates ClpB.

      Figure 3B, Supplementary Figure 5A. The solid lines from the model fit cannot be distinguished from the data points. Please modify the figures' format to clearly show the fits and the data points.

      Done.

      Lines 326, 329. It is not clear why the authors mention a lack of covalent modification of substrates by ClpB. AAA+ ATPases do not produce covalent modifications of their substrates.

      The issue of covalent modification was presented in the introduction lines 55 – 60 pointing out that much of what we have learned about protein unfolding and translocation catalyzed by ClpA and ClpX is from the observations of proteolytic degradation catalyzed by the associated protease ClpP.  However, this approach is not possible for ClpB/Hsp104 as these motors do not associate with a protease unless they have been artificially engineered to do so. 

      Lines 396-399. I am puzzled why the authors try to correlate the size of the detected kinetic step with the length of the ClpB channel instead of the size characteristics of the substrate.

      We are attempting to discuss/rationalize the observed large kinetic step-size which, in part, is defined by the structural properties of the enzyme as well as the size characteristics of the substrate.  We have attempted to clarify this and better discuss the properties of the substrate as well as ClpB.

      As I mentioned in the Public Review, it is essential to demonstrate that the emission increase used as the only readout of the ClpB position along the substrate is indeed caused by the proximity of ClpB to the fluorophore. One way to accomplish that would be to place the fluorophore upstream from the first I27 domain and determine if the "lag phase" in the emission enhancement disappears.

      Alexa Fluor 555 is well established to exhibit PIFE.  However, as in the response to the public review, we have included an appropriate control showing this in supplemental Fig. 1.

      Finally, the authors repetitively place their results in opposition to the study of Weibezahn et al. published in 2004 which first demonstrated substrate translocation by engineering a peptidase-associated variant of ClpB. It should be noted that the field of protein disaggregases has moved since the time of that publication from the initial "from-start-to-end" translocation model to a more nuanced picture of partial translocation of polypeptide loops with possible substrate slipping through the ClpB channel and a dynamic assembly of ClpB hexamers with possible subunit exchange, all of which may affect the kinetics in a complex way. However, the present study confirmed the "start-to-end" translocation model, albeit for a non-physiological ClpB substrate, and that is the take-home message, which should be included in the text.

      It is not clear to us that the field has “moved on” since Weibezahn et al 2004.  Their engineered construct that they term “BAP” with ClpP is still used in the field despite us reporting that proteolytic degradation is observed in the absence of ATP with that system  (Li, Weaver, Lin et al. 2015) and should, therefore, not be used to conclude processive energy driven translocation. The “partial translocation” by ClpB is also grounded in observations of partial degradation catalyzed by ClpP with BAP from the same group (Haslberger, Zdanowicz, Brand et al. 2008). It is not clear to us that the idea of subunit exchange leading to the possibility of assembly around internal sequences is being considered.  We do agree that this is an important mechanistic possibility that needs further interrogation. We agree with the reviewer, all these factors are confounding and lead to a more nuanced view of the mechanism.

      All that said, we have removed some of the opposition in the discussion.

      Reviewer #2 (Recommendations For The Authors):

      (1) It is assumed that the lag phase will be much longer than the phase in which we see a gradual increase in fluorescence, as the effect of PIFE is significant only when the enzyme is very close to the fluorophore. Particularly for RepA-titin3, the enzyme has to translocate many tens of nm before it is closer to the C-terminus fluorophore. However, in all cases, the lag time is lower or similar to the gradual increase phase (for example, Figure 3B). Could the authors explain this?

      The extent of the lag, or time zero until the signal starts to increase, is interpreted to indicate the time the motor moves from it’s initial binding site until it gets close enough to the fluorophore that PIFE starts to occur.  In our analysis we apply signal change to the last intermediate and dissociation or release of unfolded RepA-TitinX.  The increase in PIFE is not “all or nothing”.  Rather, it is starting to increase gradually.  Further, because these are ensemble measurements, and each molecule will exhibit variability in rate there is increased breadth of the peak due to ensemble averaging. 

      (2) Although the reason for differences in the peak position (for example, Figure 1E, 2B) is apparent, the reason for variations in the relative intensities has to be given or speculated.

      We have addressed the reason for the different peak heights in the revised manuscript.  It is the consequence of the fact that each substrate has slightly different fluorescent labeling efficiencies.  Thus, for each sample there is a mix of labeled and unlabeled substrates both of which will bind to ClpB but the unlabeled ClpB bound substrates do not contribute to the fluorescence signal, but will represent a binding competitor.  Thus, for low labeling efficiency there is a lower concentration of ClpB bound to fluorescent RepA-Titinx and for higher labeling efficiency there is higher concentration of ClpB bound to RepA-Titinx leading to an increased peak height.  RepA-Titin2 has the highest labeling efficiency and thus the largest peak height.

      Reviewer #3 (Recommendations For The Authors):

      The authors should make it clear that they and previous authors have used different constructs or conditions to bypass the physiological regulation of ClpB action by Hsp70 and its co-factors as mentioned above. In particular, the construct used by Avellaneda et al should be explained when they challenge the findings of those authors.

      Minor points:

      The lines fitting the experimental points are difficult or impossible to see in Figures 2B, 3B, and s5B.

      Fixed

      Typo bottom of p6 - "averge"

      Fixed

      Avellaneda, M. J., K. B. Franke, V. Sunderlikova, B. Bukau, A. Mogk and S. J. Tans (2020). "Processive extrusion of polypeptide loops by a Hsp100 disaggregase." Nature.

      Doyle, S. M., J. Shorter, M. Zolkiewski, J. R. Hoskins, S. Lindquist and S. Wickner (2007). "Asymmetric deceleration of ClpB or Hsp104 ATPase activity unleashes protein-remodeling activity." Nature structural & molecular biology 14(2): 114-122.

      Durie, C. L., E. C. Duran and A. L. Lucius (2018). "Escherichia coli DnaK Allosterically Modulates ClpB between High- and Low-Peptide Affinity States." Biochemistry 57(26): 3665-3675.

      Haslberger, T., A. Zdanowicz, I. Brand, J. Kirstein, K. Turgay, A. Mogk and B. Bukau (2008). "Protein disaggregation by the AAA+ chaperone ClpB involves partial threading of looped polypeptide segments." Nat Struct Mol Biol 15(6): 641-650.

      Li, T., J. Lin and A. L. Lucius (2015). "Examination of polypeptide substrate specificity for Escherichia coli ClpB." Proteins 83(1): 117-134.

      Li, T., C. L. Weaver, J. Lin, E. C. Duran, J. M. Miller and A. L. Lucius (2015). "Escherichia coli ClpB is a non-processive polypeptide translocase." Biochem J 470(1): 39-52.

      Miller, J. M., J. Lin, T. Li and A. L. Lucius (2013). "E. coli ClpA Catalyzed Polypeptide Translocation is Allosterically Controlled by the Protease ClpP." Journal of Molecular Biology 425(15): 2795-2812.

      Miller, J. M. and A. L. Lucius (2014). "ATP-gamma-S Competes with ATP for Binding at Domain 1 but not Domain 2 during ClpA Catalyzed Polypeptide Translocation." Biophys Chem 185: 58-69.

      Oberhauser, A. F., P. K. Hansma, M. Carrion-Vazquez and J. M. Fernandez (2001). "Stepwise unfolding of titin under force-clamp atomic force microscopy." Proc Natl Acad Sci U S A 98(2): 468-472.

      Rajendar, B. and A. L. Lucius (2010). "Molecular mechanism of polypeptide translocation catalyzed by the Escherichia coli ClpA protein translocase." J Mol Biol 399(5): 665-679.

      Weibezahn, J., P. Tessarz, C. Schlieker, R. Zahn, Z. Maglica, S. Lee, H. Zentgraf, E. U. Weber-Ban, D. A. Dougan, F. T. Tsai, A. Mogk and B. Bukau (2004). "Thermotolerance requires refolding of aggregated proteins by substrate translocation through the central pore of ClpB." Cell 119(5): 653-665.

    1. What a sad thing it is that a church member here & now an other of Salem, should be thus accused and charged

      I observe everyone agrees this is an unfortunate scene that had to happen. You can feel what it's their hearts is disappointment and potentially fear. These trials happened all over Protestant Europe, not just in Salem. Learning about the past based on this I understand that everyone had a hard time even suggesting this idea. This statement was made from emotion showing that regardless of the situation. Accusing people of witchcraft and then prosecuting them if found guilty would put a red pin within the township, leaving all people effected even after the trial is over.

    1. Reviewer #1 (Public Review):

      Summary:<br /> The authors used multiple approaches to study salt effects in liquid-liquid phase separation (LLPS). Results on both wild-type Caprin1 and mutants and on different types of salts contribute to a comprehensive understanding.

      Strengths:<br /> The main strength of this work is the thoroughness of investigation. This aspect is highlighted by the multiple approaches used in the study, and reinforced by the multiple protein variants and different salts studied.

      Weaknesses:<br /> (1) The multiple computational approaches are a strength, but they're cruder than explicit-solvent all-atom molecular dynamics (MD) simulations and may miss subtle effects of salts. In particular, all-atom MD simulations demonstrate that high salt strengthens pi-types of interactions (ref. 42 and MacAinsh et al, https://www.biorxiv.org/content/10.1101/2024.05.26.596000v3).

      (2) The paper can be improved by distilling the various results into a simple set of conclusions. By example, based on salt effects revealed by all-atom MD simulations, MacAinsh et al. presented a sequence-based predictor for classes of salt dependence. Wild-type Caprin1 fits right into the "high net charge" class, with a high net charge and a high aromatic content, showing no LLPS at 0 NaCl and an increasing tendency of LLPS with increasing NaCl. In contrast, pY-Caprin1 belongs to the "screening" class, with a high level of charged residues and showing a decreasing tendency of LLPS.

      (3) Mechanistic interpretations can be further simplified or clarified. (i) Reentrant salt effects (e.g., Fig. 4a) are reported but no simple explanation seems to have been provided. Fig. 4a,b look very similar to what has been reported as strong-attraction promotor and weak-attraction suppressor, respectively (ref. 50; see also PMC5928213 Fig. 2d,b). According to the latter two studies, the "reentrant" behavior of a strong-attraction promotor, CL- in the present case, is due to Cl-mediated attraction at low to medium [NaCl] and repulsion between Cl- ions at high salt. Do the authors agree with this explanation? If not, could they provide another simple physical explanation? (ii) The authors attributed the promotional effect of Cl- to counterion-bridged interchain contacts, based on a single instance. There is another simple explanation, i.e., neutralization of the net charge on Caprin1. The authors should analyze their simulation results to distinguish net charge neutralization and interchain bridging; see MacAinsh et al.

      (4) The authors presented ATP-Mg both as a single ion and as two separate ions; there is no explanation of which of the two versions reflects reality. When presenting ATP-Mg as a single ion, it's as though it forms a salt with Na+. I assume NaCl, ATP, and MgCl2 were used in the experiment. Why is Cl- not considered? Related to this point, it looks like ATP is just another salt ion studied and much of the Results section is on NaCl, so the emphasis of ATP ("Diverse Roles of ATP" in the title is somewhat misleading.

    1. Reviewer #3 (Public Review):

      Summary:

      Cheng, Liu, Dong, et al. demonstrate that anterior endoderm cells can arise from prechordal plate progenitors, which is suggested by pseudo time reanalysis of published scRNAseq data, pseudo time analysis of new scRNAseq data generated from Nodal-stimulated explants, live imaging from sox17:DsRed and Gsc:eGFP transgenics, fluorescent in situ hybridization, and a Cre/Lox system. Early fate mapping studies already suggested that progenitors at the dorsal margin give rise to both of these cell types (Warga) and live imaging from the Heisenberg lab (Sako 2016, Barone 2017) also pretty convincingly showed this. However, the data presented for this point are very nice, and the additional experiments in this manuscript, however, further cement this result. Though better demonstrated by previous work (Alexander 1999, Gritsman 1999, Gritsman 2000, Sako 2016, Rogers 2017, others), the manuscript suggests that high Nodal signaling is required for both cell types, and shows preliminary data that suggests that FGF signaling may also be important in their segregation. The manuscript also presents new single-cell RNAseq data from Nodal-stimulated explants with increased (lft1 KO) or decreased (ndr1 KD) Nodal signaling and multi-omic ATAC+scRNAseq data from wild-type 6 hpf embryos but draws relatively few conclusions from these data. Lastly, the manuscript presents data that SWI/SNF remodelers and Ripply1 may be involved in the anterior endoderm - prechordal plate decision, but these data are less convincing. The SWI/SNF remodeler experiments are unconvincing because the demonstration that these factors are differentially expressed or active between the two cell types is weak. The Ripply1 gain-of-function experiments are unconvincing because they are based on incredibly high overexpression of ripply1 (500 pg or 1000 pg) that generates a phenotype that is not in line with previously demonstrated overexpression studies (with phenotypes from 10-20x lower expression). Similarly, the cut-and-tag data seems low quality and like it doesn't support direct binding of ripply1 to these loci.

      In the end, this study provides new details that are likely important in the cell fate decision between the prechordal plate and anterior endoderm; however, it is unclear how Nodal signaling, FGF signaling, and elements of the gene regulatory network (including Gsc, possibly ripply1, and other factors) interact to make the decision. I suggest that this manuscript is of most interest to Nodal signaling or zebrafish germ layer patterning afficionados. While it provides new datasets and observations, it does not weave these into a convincing story to provide a major advance in our understanding of the specification of these cell types.

      Major issues:

      (1) UMAPs: There are several instances in the manuscript where UMAPs are used incorrectly as support for statements about how transcriptionally similar two populations are. UMAP is a stochastic, non-linear projection for visualization - distances in UMAP cannot be used to determine how transcriptionally similar or dissimilar two groups are. In order to make conclusions about how transcriptionally similar two populations are requires performing calculations either in the gene expression space, or in a linear dimensional reduction space (e.g. PCA, keeping in mind that this will only consider the subset of genes used as input into the PCA). Please correct or remove these instances, which include (but are not limited to):<br /> p.4 107-110<br /> p.4 112<br /> p.8 207-208<br /> p.10 273-275

      (2) Nodal and lefty manipulations: The section "Nodal-Lefty regulatory loop is needed for PP and anterior Endo fate specification" and Figure 3 do not draw any significant conclusions. This section presents a LIANA analysis to determine the signals that might be important between prechordal plate and endoderm, but despite the fact that it suggests that BMP, Nodal, FGF, and Wnt signaling might be important, the manuscript just concludes that Nodal signaling is important. Perhaps this is because the conclusion that Nodal signaling is required for the specification of these cell types has been demonstrated in zebrafish in several other studies with more convincing experiments (Alexander 1999, Gritsman 1999, Gritsman 2000, Rogers 2017, Sako 2016). While FGF has recently been demonstrated to be a key player in the stochastic decision to adopt endodermal fate in lateral endoderm (Economou 2022), the idea that FGF signaling may be a key player in the differentiation of these two cell types has strangely been relegated to the discussion and supplement. Lastly, the manuscript does not make clear the advantage of performing experiments to explore the PP-Endo decision in Nodal-stimulated explants compared to data from intact embryos. What would be learned from this and not from an embryo? Since Nodal signaling stimulates the expression of Wnts and FGFs, these data do not test Nodal signaling independent of the other pathways. It is unclear why this artificial system that has some disadvantages is used since the manuscript does not make clear any advantages that it might have had.

      (3) ripply1 mRNA injection phenotype inconsistent with previous literature: The phenotype presented in this manuscript from overexpressing ripply1 mRNA (Fig S11) is inconsistent with previous observations. This study shows a much more dramatic phenotype, suggesting that the overexpression may be to a non-physiological level that makes it difficult to interpret the gain-of-function experiments. For instance, Kawamura et al 2005 perform this experiment but do not trigger loss of head and eye structures or loss of tail structures. Similarly, Kawamura et al 2008 repeat the experiment, triggering a mildly more dramatic shortening of the tail and complete removal of the notochord, but again no disturbance of head structures as displayed here. These previous studies injected 25 - 100 pg of ripply1 mRNA with dramatic phenotypes, whereas this study uses 500 - 1000 pg. The phenotype is so much more dramatic than previously presented that it suggests that the level of ripply1 overexpression is sufficiently high that it may no longer be regulating only its endogenous targets, making the results drawn from ripply1 overexpression difficult to trust.

      (4) Ripply1 binding to sox17 and sox32 regulatory regions not convincing: The Cut and Tag data presented in Fig 6J-K does not seem to be high quality and does not seem to provide strong support that Ripply 1 binds to the regulatory regions of these genes. The signal-to-noise ratio is very poor, and the 'binding' near sox17 that is identified seems to be even coverage over a 14 kb region, which is not consistent with site-specific recruitment of this factor, and the 'peaks' highlighted with yellow boxes do not appear to be peaks at all. To me, it seems this probably represents either: (1) overtagmentation of these samples or (2) an overexpression artifact from injection of too high concentration of ripply1-HA mRNA. In general, Cut and Tag is only recommended for histone modifications, and Cut and Run would be recommended for transcriptional regulators like these (see Epicypher's literature). Given this and the previous point about Ripply1 overexpression, I am not convinced that Ripply1 regulates endodermal genes. The existing data could be made somewhat more convincing by showing the tracks for other genes as positive and negative controls, given that Ripply1 has known muscle targets (how does its binding look at those targets in comparison) and there should be a number of Nodal target genes that Ripply1 does not bind to that could be used as negative controls. Overall this experiment doesn't seem to be of high enough quality to drive the conclusion that Ripply1 directly binds near sox17 and sox32 and from the data presented in the manuscript looks as if it failed technically.

      (5) "Cooperatively Gsc and ripply1 regulate": I suggest avoiding the term "cooperative," when describing the relationship between Ripply1 and Gsc regulation of PP and anterior endoderm - it evokes the concept of cooperative gene regulation, which implies that these factors interact with each biochemically in order to bind to the DNA. This is not supported by the data in this manuscript, and is especially confusing since Ripply1 is thought to require cooperative binding with a T-box family transcription factor to direct its binding to the DNA.

      (6) SWI/SNF: The differential expression of srcap doesn't seem very remarkable. The dot plots in the supplement S7H don't help - they seem to show no expression at all in the endoderm, which is clearly a distortion of the data, since from the violin plots it's obviously expressed and the dot-size scale only ranges from ~30-38%. Please add to the figure information about fold-change and p-value for the differential expression. Publicly available scRNAseq databases show scrap is expressed throughout the entire early embryo, suggesting that it would be surprising for it to have differential activity in these two cell types and thereby contribute to their separate specification during development. It seems equally possible that this just mildly influences the level of Nodal or FGF signaling, which would create this effect.

      The multiome data seems like a valuable data set for researchers interested in this stage of zebrafish development. However, the presentation of the data doesn't make many conclusions, aside from identifying an element adjacent to ripply1 whose chromatin is open in prechordal plate cells and not endodermal cells and showing that there are a number of loci with differential accessibility between these cell types. That seems fairly expected since both cell types have several differentially expressed transcriptional regulators (for instance, ripply1 has previously been demonstrated in multiple studies to be specific to the prechordal plate during blastula stages). The manuscript implies that SWI/SNF remodeling by Srcap is responsible for the chromatin accessibility differences between these cell types, but that has not actually been tested. It seems more likely that the differences in chromatin accessibility observed are a result of transcription factors binding downstream of Nodal signaling.

      Minor issues:

      Figure 2 E-F: It's not clear which cells from E are quantitated in F. For instance, the dorsal forerunner cells are likely to behave very differently from other endodermal progenitors in this assay. It would be helpful to indicate which cells are analyzed in Fig F with an outline or other indicator of some kind. Or - if both DFCs and endodermal cells are included in F, to perhaps use different colors for their points to help indicate if their fluorescence changes differently.

      Fig 3 J: Should the reference be Dubrulle et al 2015, rather than Julien et al?

      References:<br /> Alexander, J. & Stainier, D. Y. A molecular pathway leading to endoderm formation in zebrafish. Current biology : CB 9, 1147-1157 (1999).<br /> Barone, V. et al. An Effective Feedback Loop between Cell-Cell Contact Duration and Morphogen Signaling Determines Cell Fate. Dev. Cell 43, 198-211.e12 (2017).<br /> Economou, A. D., Guglielmi, L., East, P. & Hill, C. S. Nodal signaling establishes a competency window for stochastic cell fate switching. Dev. Cell 57, 2604-2622.e5 (2022).<br /> Gritsman, K. et al. The EGF-CFC protein one-eyed pinhead is essential for nodal signaling. Cell 97, 121-132 (1999).<br /> Gritsman, K., Talbot, W. S. & Schier, A. F. Nodal signaling patterns the organizer. Development (Cambridge, England) 127, 921-932 (2000).<br /> Kawamura, A. et al. Groucho-associated transcriptional repressor ripply1 is required for proper transition from the presomitic mesoderm to somites. Developmental cell 9, 735-744 (2005).<br /> Kawamura, A., Koshida, S. & Takada, S. Activator-to-repressor conversion of T-box transcription factors by the Ripply family of Groucho/TLE-associated mediators. Molecular and cellular biology 28, 3236-3244 (2008).<br /> Sako, K. et al. Optogenetic Control of Nodal Signaling Reveals a Temporal Pattern of Nodal Signaling Regulating Cell Fate Specification during Gastrulation. Cell Rep. 16, 866-877 (2016).<br /> Rogers, K. W. et al. Nodal patterning without Lefty inhibitory feedback is functional but fragile. eLife 6, e28785 (2017).<br /> Warga, R. M. & Nüsslein-Volhard, C. Origin and development of the zebrafish endoderm. Development 126, 827-838 (1999).

    1. Hey, you made it all the way down here, so before you go, why not take a chance on making art for yourself.  Enjoy theBring some joy into your life! No risk, no obligation, just a chance to make something with your hands! And it's FREE...

      If you're after a fun creative project that you ACTUALLY take from idea to custom art journey, then get FREE instant access now.

      And I'd make the button somewhat less doubtful ;-) SIGN ME UP NOW

    1. students sing the ABCs as a means to other ends—remembering the letters and sequence of the alphabet

      I never thought of this as a way that arts help learning! It just felt like such an easy and obvious way of memorizing the alphabet. It's incredible ow much art can enhance learning, and I look forward of discovering and creating new ways of applying this method of teaching in my future classroom.

    1. Performance measures for different tree reconstruction method. a) Kuhner-Felsenstein (KF) distance, which takes into account both topology and branch lengths of the compared trees; b) mean absolute error (MAE) on pairwise distances, which ignores topology; c) normalized Robinson-Foulds (RF) distance, which only takes into account tree topology. The alignments for which trees are inferred, were simulated under the LG+GC sequence model and are all 500 amino acids long. For each measure, we show 95% confidence intervals estimated with 1000 bootstrap samples.

      These results paired with the runtimes are really quite impressive! But the contrast between the results for KF distances as compared to RF distances are interesting, and seems like they may be worth unpacking.

      In particular, it's notable that the RF distances at greater tree sizes for PF+FastME seem to converge with FastME, being greater than seen for IQTree/FastTree, with the difference increasing along with tree size.

      As you say, RF is just the sum of differences in bipartitions between two trees, whereas KF considers both differences in topology and branch length. You find that PF+FastME consistently infers trees with lower or equivalent KF distances to IQTree and FastTree. But, as tree size increases, RF distances increase for PF+FastME at a high rate, exceeding those of FastTree and IQTree starting at relatively small trees (~20 tips).

      Together, these results would suggest that PF+FastME estimates branch lengths well. This is maybe expected but a great thing to see, since PF is effectively trained to infer those evolutionary distances that FastME uses to infer branch lengths! However, despite accurately inferring branch-lengths, there seems to be a larger number of topological errors in the larger trees inferred by PF+FastME as compared to the other methods.

      Do you have any intuition as to why this discrepancy arises? Or any thoughts on how you might modify the model/model architecture to better account for and mitigate this effect?

    1. The sightseer may be aware that something is wrong. He maysimply be bored; or he may be conscious of the difficulty: that thegreat thing yawning at his feet somehow eludes him. The harder helooks at it, the less he can see.

      Funny how a natural wonder, in some mental image sense, is only as beautiful and marvelous as society accumulates it to be. postcards, paintings, articles, etc set a general expectation, turning a visit to see said natural wonder into a visit purely just to visit, and our thoughts at the sight of it are inseparable from another person's description of it's beauty. Like a blind ritual of a sort, something we people just do to do it. To go and see and have some life changing, awe-inspiring moment because we simply must. And then imagine it's a rainy day and the sight of whatever natural wonder is ruined. Maybe we'd want to be disappointed. Maybe we wouldn't allow ourslves to be.

    1. it is the only instrument

      I feel like it's a little alienating to say the only instrument. We don't have kazoos in orchestras (usually) and I don't think it would hurt to just say that it's widely excluded from symphonies and orchestras.

    1. Her response: It’s her job to ensure students develop basic writing skills, and the noticeable uptick in AI use is impeding those efforts

      This must be very frustrating for teachers who truly care about the students and their progression, but they are just trying to get through another class. I don't think I ever really thought about instructors in this scenario because I feel like I haven't ever really had a teacher truly passionate enough about me and my education. That has made it a bit harder for me to actually put my best effort in their classes because if they aren't invested in me then why should I be invested in their class. This is mostly for all of my high school classes, but could be applicable to higher education as well.

    2. A few didn’t know they had used generative AI because it’s embedded in so many other tools, like Grammarly.

      This is interesting, I didn't know that Grammarly used generative AI. Now that I think about it it makes sense, but it is a bit unnerving that we could be using AI for assignments and not even know about it.

    3. A few didn’t know they had used generative AI because it’s embedded in so many other tools, like Grammarly

      Are there laws that require businesses to state up-front that they are using AI? If not, should students still be held accountable?

    1. indigo blue

      A very strong tie-in to the word choice for the book title. Again, it's blue jeans, not just jeans. This is also an excellent indicator that the book will globalize its discussion, which I commented on earlier with regard to its specifically American bent.

    1. One can still enjoy life’s simple pleasures–friendship, celebration, entertainment, family– just like non-philosophers, but in addition to these, you will develop the deeper joys of contemplation, understanding, and personal commitment.

      It can be lonely at times and sometimes have you questioning your own sanity, but it's a journey that every person is walking at their own pace in their own time. With awareness or without it.

    1. democratization of publishing

      worth noting, that its not just publishing, but mass media and transmissions, not unlike the global village as mccluhan predicted. its in every area of life.

      the gatekeeping previously at least had certain checks and balances, and its part of why platforms like tiktok and the platform's spread of misinformation has become such a lightning rod issue

      on the other hand UMG for example, had to make a deal though to, admitting that the platform had created and validated a lucrative market

      and that's the thing.

      Blurred lines. Music, news, sports, and in every aspect of life, there's grey. Do you want to be positive or not?

      UMG, a public company, has a publishing division. They publish music, and they often have to work with newsrooms to license out music, if say they want a particular song as ESPN might, for example - instead of "library music" -- which is another type of company, that ultimately, functions, as a publisher.

      Linda Villarosa, is quite the accomplished multi-hyphenate, and that includes publisher. It's not the first thing she discloses or usually is referred to as, but she has a publishing company and she published her own novel under it. Whether or not she partnered with a University Press or a for-profit company, or printed and bounded it all by herself, her books, were published via Linda's publishing and she is a publisher

    1. Utilization effectiveness

      This is a new term to me. I think it's bit like PUE in that you want the number to be as close to 1 as possible (i.e. for 1 server's worth of utilisation, you want to have provision 1 server, not 5)

      So, I think this is the term you might use to talk about how a hyperscaler datacentre might be full of individually efficient servers, but only have 5% of them serving workloads - the rest is just spare capacity ready to sell, or give away as credits

    1. f it looks just like the postcard

      Some people only look at nature for the look or it, or for the postcard view. I think that nature should be appreciated for what it is. I think it's good to add this in the story though, because it shows that side of thinking.

    1. 7. ‘It Was a Mistake’This September, with late-season heat pounding Washington, Zenzi Suhadi cleared security at the Russell SenateOffice Building, preparing to brief Senate aides about the impact that palm-oil development was having onIndonesia’s environment. I asked if he was nervous, and he said no. “These are just people. I don’t have to face anytigers.” He didn’t seem to be joking.Suhadi wanted to tell the lawmakers the same thing he told them in two previous visits to Capitol Hill: that thepalm trade, driven by American investment, is slowly killing his country. “It’s important for you to understand thatall acts of deforestation in Indonesia start with a signature,” he said. “And more than a little of it starts right here.”He was not confident that he would be heard —the last time he visited Washington, lawmakers chewed up theirtime asking him about water buffalo in his village. But still he felt compelled to speak.From Washington, Suhadi traveled to San Francisco to attend a climate march and address a group of hedge-fundinvestors. Just down the street, Michael Bloomberg implored strong immediate action on emissions reductions atthe Global Climate Action Summit, one of the nation’s largest gatherings on climate goals. But the conference waslight on substance when it came to the subject of forests. There was scarcely any mention of peatland at all.PDF GENERATED BY PROQUEST.COM Page 10 of 12

      economy was shifted cause of it

    1. One weakness in Plato's position was that, to make his objections effective, he put theminto writing, just as one weakness in anti-print positions is that their proponents, to make theirobjections more effective, put the objections into print.

      Lol, it's quite literally a cycle. That or writing, print, spoken word, and computers are intrinsically linked. It's impossible to leverage one over the other, as all are needed in a constantly shifting society.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      In this paper, the researchers aimed to address whether bees causally understand string-pulling through a series of experiments. I first briefly summarize what they did:

      - In experiment 1, the researchers trained bees without string and then presented them with flowers in the test phase that either had connected or disconnected strings, to determine what their preference was without any training. Bees did not show any preference.

      - In experiment 2, bees were trained to have experience with string and then tested on their choice between connected vs. disconnected string.

      - experiment 3 was similar except that instead of having one option which was an attached string broken in the middle, the string was completely disconnected from the flower.

      - In experiment 4, bees were trained on green strings and tested on white strings to determine if they generalize across color.

      - In experiment 5, bees were trained on blue strings and tested on white strings.

      - In experiment 6, bees were trained where black tape covered the area between the string and the flower (i.e. so they would not be able to see/ learn whether it was connected or disconnected).

      - In experiments 2-6, bees chose the connected string in the test phase.

      - In experiment 7, bees were trained as in experiment 3 and then tested where the string was either disconnected or coiled i.e. still being 'functional' but appearing different.

      - In experiment 8, bees were trained as before and then tested on a string that was in a different coiled orientation, either connected or disconnected.

      - In experiments 7 and 8 the bees showed no preference.

      Strengths:

      I appreciate the amount of work that has gone into this study and think it contains a nice, thorough set of experiments. I enjoyed reading the paper and felt that overall it was well-written and clear. I think experiment 1 shows that bees do not have an untrained understanding of the function of the string in this context. The rest of the experiments indicate that with training, bees have a preference for unbroken over broken string and likely use visual cues learned during training to make this choice. They also show that as in other contexts, bees readily generalize across different colors.

      Weaknesses:

      (1) I think there are 2 key pieces of information that can be taken from the test phase - the bees' first choice and then their behavior across the whole test. I think the first choice is critical in terms of what the bee has learned from the training phase - then their behavior from this point is informed by the feedback they obtain during the test phase. I think both pieces of information are worth considering, but their behavior across the entire test phase is giving different information than their first choice, and this distinction could be made more explicit. In addition, while the bees' first choice is reported, no statistics are presented for their preferences.

      We agree with the reviewer that the first choice is critical in terms of what the bumblebees have learned from the training phase. We analyzed the bees’ first choice in Table 1, and we added the tested videos. The entire connected and disconnected strings were glued to the floor, the bees were unable to move either the connected or disconnected strings, and avoid learning behavior during the tests. We added the data of bee's each choice in the Supplementary table.

      (2) It seemed to me that the bees might not only be using visual feedback but also motor feedback. This would not explain their behavior in the first test choice, but could explain some of their subsequent behavior. For example, bees might learn during training that there is some friction/weight associated with pulling the string, but in cases where the string is separated from the flower, this would presumably feel different to the bee in terms of the physical feedback it is receiving. I'd be interested to see some of these test videos (perhaps these could be shared as supplementary material, in addition to the training videos already uploaded), to see what the bees' behavior looks like after they attempt to pull a disconnected string.

      We added supplementary videos of testing phase. As noted in General Methods, both connected and disconnected strings were glued to the floor to prevent the air flow generated by flying bumblebees’ wings from changing the position of the string during the testing phase. The bees were unable to move either the connected or disconnected strings during the tests, and only attempted to pull them. Therefore, the difference in the friction/weight of pulling the both strings cannot be a factor in the test.

      (3) I think the statistics section needs to be made clearer (more in private comments).

      We changed the statistical analysis section as suggested by the reviewer.

      (4) I think the paper would be made stronger by considering the natural context in which the bee performs this behavior. Bees manipulate flowers in all kinds of contexts and scrabble with their legs to achieve nectar rewards. Rather than thinking that it is pulling a string, my guess would be that the bee learns that a particular motor pattern within their usual foraging repertoire (scrabbling with legs), leads to a reward. I don't think this makes the behavior any less interesting - in fact, I think considering the behavior through an ecological lens can help make better sense of it.

      Here we respectfully disagree. The solving of Rubik’s cube by humans could be said to be version of finger-movements naturally required to open nuts or remove ticks from fur, but this is somewhat beside the point: it’s not the motor sequences that are of interest, but the cognition involved. A general approach in work on animal intelligence and cognition is to deliberately choose paradigms that are outside the animals’ daily routines-this is what we have done here, in asking whether there is means-end comprehension in bee problem solving. Like comparable studies on this question in other animals, the experiments are designed to probe this question, not one of ecological validity.

      Reviewer #2 (Public Review):

      Summary:

      The authors wanted to see if bumblebees could succeed in the string-pulling paradigm with broken strings. They found that bumblebees can learn to pull strings and that they have a preference to pull on intact strings vs broken ones. The authors conclude that bumblebees use image matching to complete the string-pulling task.

      Strengths:

      The study has an excellent experimental design and contributes to our understanding of what information bumblebees use to solve a string-pulling task.

      Weaknesses:

      Overall, I think the manuscript is good, but it is missing some context. Why do bumblebees rely on image matching rather than causal reasoning? Could it have something to do with their ecology? And how is the task relevant for bumblebees in the wild? Does the test translate to any real-life situations? Is pulling a natural behaviour that bees do? Does image matching have adaptive significance?

      We appreciate the valuable comment from the reviewer. Our explanation, which we have now added to the manuscript, is as follows:

      “Different flower species offer varying profitability in terms of nectar and pollen to bumblebees; they need to make careful choices and learn to use floral cues to predict rewards (Chittka, 2017). Bumblebees can easily learn visual patterns and shapes of flower (Meyer-Rochow, 2019); they can detect stimuli and discriminate between differently coloured stimuli when presented as briefly as 25 ms (Nityananda et al., 2014). In contrast, causal reasoning involves understanding and responding to causal relationships. Bumblebees might favor, or be limited to, a visual approach, likely due to the efficiency and simplicity of processing visual cues to solve the string-pulling task. ”

      As above, it worth noting that our work is not designed as an ecological study, but one about the question of whether causal reasoning can explain how bees solve a string-pulling puzzle. We have a cognitive focus, in line with comparable studies on other animals. We deliberately chose a paradigm that is to some extent outside of the daily challenges of the animal.

      Reviewer #3 (Public Review):

      Summary:

      This paper presents bees with varying levels of experience with a choice task where bees have to choose to pull either a connected or unconnected string, each attached to a yellow flower containing sugar water. Bees without experience of string pulling did not choose the connected string above chance (experiment 1), but with experience of horizontal string pulling (as in the right-hand panel of Figure 4) bees did choose the connected string above chance (experiments 2-3), even when the string colour changed between training and test (experiments 4-5). Bees that were not provided with perceptual-motor feedback (i.e they could not observe that each pull of the string moved the flower) during training still learned to string pull and then chose the connected string option above chance (experiment 6). Bees with normal experience of string pulling then failed to discriminate between connected and unconnected strings when the strings were coiled or looped, rather than presented straight (experiments 7-8).

      Weaknesses:

      The authors have only provided video of some of the conditions where the bees succeeded. In general, I think a video explaining each condition and then showing a clip of a typical performance would make it much easier to follow the study designs for scholars. Videos of the conditions bees failed at would be highly useful in order to compare different hypotheses for how the bees are solving this problem. I also think it is highly important to code the videos for switching behaviours. When solving the connected vs unconnected string tasks, when bees were observed pulling the unconnected string, did they quickly switch to the other string? Or did they continue to pull the wrong string? This would help discriminate the use of perceptual-motor feedback from other hypotheses.

      We added the test videos as suggested by the reviewer, and we added the data for each bee's choice. However, both connected and disconnected strings were glued to the floor, and therefore perceptual-motor feedback was equal and irrelevant between the choices during the test.

      The experiments are also not described well, for my below comments I have assumed that different groups of bees were tested for experiments 1-8, and that experiment 6 was run as described in line 331, where bees were given string-pulling training without perceptual feedback rather than how it is described in Figure 4B, which describes bees as receiving string pulling training with feedback.

      We now added figures of Experiment 6 and 7 in the Figure 1B, and we mentioned that different groups of bees were tested for Experiments 1-9.

      The authors suggest the bees' performance is best explained by what they term 'image matching'. However, experiment 6 does not seem to support this without assuming retroactive image matching after the problem is solved. The logic of experiment 6 is described as "This was to ensure that the bees could not see the familiar "lollipop shape" while pulling strings....If the bees prefer to pull the connected strings, this would indicate that bees memorize the arrangement of strings-connected flowers in this task." I disagree with this second sentence, removing perceptual feedback during training would prevent bees memorising the lollipop shape, because, while solving the task, they don't actually see a string connected to a yellow flower, due to the black barrier. At the end of the task, the string is now behind the bee, so unless the bee is turning around and encoding this object retrospectively as the image to match, it seems hard to imagine how the bee learns the lollipop shape.

      We agree with the reviewer that while solving the task in the last step during training, the bees don't actually see a string connected to a yellow flower, due to the black barrier. Since the full shape is only visible after the pulling is completed and this requires the bee to “check back” on the entire display after feeding, to basically conclude “ this is the shape that I need to be looking for later”.

      Another possibility is that bumblebees might remember the image of the “lollipop shape” while training the bees in the first step, in which the “lollipop shape” was directly presented to the bumblebee in the early step of the training.

      We added the experiment suggested by the reviewer, and the result showed that when a green table was placed behind the string to obscure the “lollipop shape” at any point during the training phase, the bees were unable to identify the connected string. The result further supports that bumblebees learn to choose the connected string through image matching.

      Despite this, the authors go on to describe image matching as one of their main findings. For this claim, I would suggest the authors run another experiment, identical to experiment 6 but with a black panel behind the bee, such that the string the bee pulls behind itself disappears from view. There is now no image to match at any point from the bee's perspective so it should now fail the connectivity task.

      Strengths:

      Despite these issues, this is a fascinating dataset. Experiments 1 and 2 show that the bees are not learning to discriminate between connected and unconnected stimuli rapidly in the first trials of the test. Instead, it is clear that experience in string pulling is needed to discriminate between connected and unconnected strings. What aspect of this experience is important? Experiment 6 suggests it is not image matching (when no image is provided during problem-solving, but only afterward, bees still attend to string connectivity) and casts doubt on perceptual-motor feedback (unless from the bee's perspective, they do actually get feedback that pulling the string moves the flower, video is needed here). Experiments 7 and 8 rule out means-end understanding because if the bees are capable of imagining the effect of their actions on the string and then planning out their actions (as hypotheses such as insight, means-end understanding and string connectivity suggest), they should solve these tasks. If the authors can compare the bees' performance in a more detailed way to other species, and run the experiment suggested, this will be a highly exciting paper

      We appreciate the valuable comment from the reviewer. We compared the bees' performance to other species, and conducted the experiment as suggested by the reviewer.

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors):

      Smaller comments:

      Line 64: is the word 'simple' needed here? It could also be explained by more complex forms of associative learning, no?

      We deleted “simple”.

      Methods:

      Line 230: was it checked that this was high-contrast for the bees?

      We added the relevant reference in the revised manuscript.

      Line 240: how much sucrose solution was present in the flowers?

      We added 25 microliters sucrose solution in the flowers. We added the information in the revised manuscript.

      Line 266: check grammar.

      We checked the grammar as follows: “During tests, both strings were glued to the floor of the arena to prevent the air flow generated by flying bumblebees’ wings from changing the position of the string.”

      Statistical analysis:

      - What does it mean that "Bees identity and colony were analyzed with likelihood ratio tests"?

      Bees identity and colony was set as a random variable. We changed the analysis methods in the revised manuscript, and results of the all the experiments did not changed.

      - Line 359: do you mean proportion rather than percentage?

      We mean the percentage.

      - "the number of total choices as weights" - this should be explained further. This is the number of choices that each bee made? What was the variation and mean of this number? If bees varied a lot in this metric, it might make more sense to analyze their first choice (as I see you've done) and their first 10 choices or something like that - for consistency.

      This refers to the total number of choices made by each bumblebee. We added the mean and standard error of each bee’s number of choices in Table 1. Some bees pulled the string fewer than 10 times; we chose to include all choices made by each bee.

      - More generally I think the first test is more informative than the subsequent choices, since every choice after their first could be affected by feedback they are getting in that test phase. Or rather, they are telling you different things.

      All the bees were tested only once, however, you might be referring to the first choice. We used Chi-square test to analyze the bumblebees’ first choices in the test. It is worth noting that both connected and disconnected strings were glued to the floor. The bees were unable to move either the connected or disconnected strings during the tests, and only attempted to pull them. Therefore,the feedback from pulling either the connected or disconnected strings is the same.

      - Line 362: I think I know what you mean, but this should be re-phrased because the "number of" sounds more appropriate for a Poisson distribution. I think what you are testing is whether each individual bee chose the connected or the disconnected string - i.e. a 0 or 1 response for each bee?

      We agree with the reviewer that each bee chose the connected or the disconnected string - i.e. a 0 or 1 response for each bee, but not the number. We clarify this as: “The total number of the choices made by each bee was set as weights.” 

      - Line 364-365: here and elsewhere, every time you mention a model, make it clear what the dependent and independent variables are. i.e. for the mixed model, the 'bee' is the random factor? Or also the colony that the bee came from? Were these nested etc?

      We clarify this in the revised manuscript. The bee identity and colony is the random factor in the mixed model.

      - Line 368: "Latency to the first choice of each bee was recorded" - why? What were the hypotheses/ predictions here?

      The latency to the first choice was intended to see if the bumblebees were familiarizing with the testing pattern. A shorter delay time might indicate that the bumblebees were more familiar with the pattern.

      - Line 371: "Multiple comparisons among experiments were.." - do you mean 'within' experiments? It seems that treatments should not be compared between different experiments.

      We mean multiple comparisons among different experiments; we clarify this in the revised manuscript.

      Results

      Experiment 1: From the methods, it sounded like you both analyzed the bees' first choice and their total no. of choices, but in the results section (and Figure 1) I only see the data for all choices combined here.

      In table 1 and in the text you report the number of bees that chose each option on their first choice, but there are no statistical results associated with these results. At the very least, a chi square or binomial test could be run.

      Line 138: "Interestingly, ten out of fifteen bees pulled the connected string in their first choice" - this is presented like it is a significant majority of bees, but a chi-square test of 10 vs 5 has a p-value = 0.1967

      We used the Chi square test to analyzed of the bees’ first choice. We also added the analyzed data in the Table 1.

      Line 143: "It makes sense because the bees could see the "lollipop shape" once they pulled it out from the table." - this feels more like interpretation (i.e. Discussion) rather than results.

      We moved the sentence to the discussion.

      Line 162: again this feels more like interpretation/ conjecture than results.

      We removed the sentence in the results.

      Line 184: check grammar.

      We checked the grammar. We changed “task” to “tasks”.

      Figures

      I really appreciated the overview in Figure 5 - though I think this should be Figure 1? Even if the methods come later in eLife, I think it would be nice to have that cited earlier on (e.g. at the start of the results) to draw the reader's attention to it quickly, since it's so helpful. It also then makes the images at the bottom of what is currently Figure 1 make more sense. I also think that the authors could make it clearer in Figure 5 which strings are connected vs disconnected in the figure (even if it means exaggerating the distance more than it was in real life). I had to zoom in quite a bit to see which were connected vs. not. Alternatively, you could have an arrow to the string with the words "connected" "disconnected" the first time you draw it - and similar labels for the other string conditions.

      We appreciate the valuable comment from the reviewer. We changed Figure 5 to Figure 2, and Figure 4 to Figure 1. We cited the Figures at the start of the results. We also changed the gap distance between the disconnected strings. Additionally, we added arrows to indicate “connected” and “disconnected” strings in the Figure.

      Figure 1 - I think you could make it clearer that the bars refer to experiments (e.g. have an x-axis with this as a label). Also, check the grammar of the y-axis.

      We added the experiments number in the Figures. Additionally, we checked the grammar of the y-axis. We changed “percentages” to “parentage”. 

      I also think it's really helpful to see the supplementary videos but I think it would be nice to see some examples of the test phase, and not just the training examples.

      We added Supplementary videos of the testing phase.

      Reviewer #2 (Recommendations For The Authors):

      Below are also some minor comments:

      L40: "approaches".

      We changed “approach” to “approaches”.

      L42: but likely mainly due to sampling bias of mammals and birds.

      We changed the sentence as follows: String pulling is one of the most extensively used approaches in comparative psychology to evaluate the understanding of causal relationships (Jacobs & Osvath, 2015), with most research focused on mammals and birds, where a food item is visible to the animal but accessible only by pulling on a string attached to the reward (Taylor, 2010; Range et al., 2012; Jacobs & Osvath, 2015; Wakonig et al., 2021).

      L64: remove "in this study"

      We removed “in this study”.

      L64: simple associative learning of what? Isn't your image matching associative too?

      We removed “ simple”.

      L97: remove "a" before "connected".

      We removed “a” before “connected”.

      L136-138: but maybe they could still feel the weight of the flower when pulling?

      Because both strings were glued to the floor in the test phase, the feedback was the same and therefore irrelevant. This information is noted in the General Methods.

      L161: what are these numbers?

      We removed the latency in the revised manuscript.

      L167/ Table 1: I realise that the authors never tried slanted strings to check if bumblebees used proximity as a cue. Why?

      This was simply because we wanted to focus on whether bumblebees could recognize the connectivity of the string.

      Discussion: Why did you only control for colour of the string? What if you had used strings with different textures or smells? Unclear if the authors controlled for "bumblebee smell" on the strings, i.e., after a bee had used the string, was the string replaced by a new one or was the same one used multiple times?

      We used different colors to investigate featural generalization of the visual display of the string connected to the flower in this task. We controlled for color because it is a feature that bumblebees can easily distinguish.

      Both the flowers and the strings were used only once, to prevent the use of chemosensory cues. We clarify this in the revised manuscript.

      L182: since what?

      We deleted “since” in the revised manuscript.

      L182-188: might be worth mentioning that some crows and parrots known for complex cognition perform poorly on broken strings (e.g., https://doi.org/10.1098/rspb.2012.1998 ; https://doi.org/10.1163/1568539X-00003511 ; https://doi.org/10.1038/s41598-021-94879-x ) and Australian magpies use trial and error (https://doi.org/10.1007/s00265-023-03326-6).

      We added the following sentences as suggested by the reviewer: “It is worth noting that some crows and parrots known for complex cognition perform poorly on the broken string task without perceptual feedback or learning. For example, New Caledonian crows use perceptual feedback strategies to solve the broken string-pulling task, and no individual showed a significant preference for the connected string when perceptual feedback was restricted (Taylor et al., 2012). Some Australian magpies and African grey parrots can solve the broken string task, but they required a high number of trials, indicating that learning plays a crucial role in solving this task (Molina et al., 2019; Johnsson et al., 2023).”

      L193: maybe expand on this to put the task into a natural context?

      We added the following sentences as suggested by the reviewer:

      “Different flower species offer varying profitability in terms of nectar and pollen to bumblebees; they need to make careful choices and learn to use floral cues to predict rewards (Chittka, 2017). Bumblebees can easily learn visual patterns and shapes of flower (Meyer-Rochow, 2019); they can detect stimuli and discriminate between differently coloured stimuli when presented as briefly as 25 ms (Nityananda et al., 2014). In contrast, causal reasoning involves understanding and responding to causal relationships. Bumblebees might favor, or be limited to, a visual approach, likely due to the efficiency and simplicity of processing visual cues to solve the string-pulling task. ”

      L204: is causal understanding the same as means-end understanding?

      Means-end understanding is expressed as goal-directed behavior, which involves the deliberate and planned execution of a sequence of steps to achieve a goal. Includes some understanding of the causal relationship (Jacobs & Osvath, 2015; Ortiz et al., 2019). .

      L235: this is a very big span of time. Why not control for motivation? Cognitive performance can vary significantly across the day (at least in humans).

      Bumblebee motivation is understood to be rather consistent, as those that were trained and tested came to the flight arena of their own volition and were foragers looking to fill their crop load each time to return it to the colony.

      L232: what is "(w/w)" ? This occurs throughout the manuscript.

      “w/w” represents the weight-to-weight percentage of sugar.

      L250: this sentence sounds odd. "containing in the central well.." ?? Perhaps rephrase? Unclear what central well refers to? Did the flowers have multiple wells?

      We rephrased the sentence as follows: For each experiment, bumblebees were trained to retrieve a flower with an inverted Eppendorf cap at the center, containing 25 microliters of 50% sucrose solution, from underneath a transparent acrylic table

      L268: why euthanise?

      The reason for euthanizing the bees is that new foragers will typically only become active after the current ones were removed from the hive.

      L270: chemosensory cues answer my concern above. Maybe make it clear earlier.

      We moved this sentence earlier in the result.

      L273: did different individuals use different pulling strategies? Do you have the data to analyse this? This has been done on birds and would offer a nice comparison.

      We analyzed the string-pulling strategies among different individuals, and provided Supplementary Table 1 to display the performances of each individual in different string-pulling experiments.

      L365: unclear why both models. Would be nice to see a GLM output table.

      The duration of pulling different kinds of strings were first tested with the Shapiro-Wilk test to assess data normality. The duration data that conforms to a normal distribution was compared using linear mixed-effects models (LMM), while the data that deviates from normality were examined with a generalized linear-mixed model (GLMM). We added a GLM and GLMM output table in the revised manuscript.

      L377: should be a space between the "." and "This".

      We added a space between the “.” and “This”.

      L383-390: some commas and semicolons are in the wrong places.

      We carefully checked the commas and semicolons in this sentence.

      Reviewer #3 (Recommendations For The Authors):

      Minor comments

      Line 32: seems to be missing a word, suggest "the bumblebees' ability to distinguish".

      we added “the” in the revised manuscript.

      Line 47: it would be good to reference other scholars here, this is the central focus of all work in comparative psychology.

      We added the reference in the revised manuscript.

      Line 50-61: I think the string-pulling literature could be described in more detail here, with mention of perceptual-motor feedback loops as a competing hypothesis to means-end understanding (see Taylor et al 2010, 2012). It seems a stretch to suggest that "String-pulling studies have directly tested means-end comprehension in various species", when perceptual-motor feedback is a competing hypothesis that we have positive evidence for in several species.

      We mentioned the perceptual-motor feedback in the introduction as follow:

      “Multiple mechanisms can be involved in the string-pulling task, including the proximity principle, perceptual feedback and means-end understanding (Taylor et al., 2012; Wasserman et al., 2013; Jacobs & Osvath, 2015; Wang et al., 2020). The principle of proximity refers to animals preferring to pull the reward that is closest to them (Jacobs & Osvath, 2015). Taylor et al. (2012) proposed that the success of New Caledonian crows in string-pulling tasks is based on a perceptual-motor feedback loop, where the reward gradually moves closer to the animal as they pull the strings. If the visual signal of the reward approaching is restricted, crows with no prior string-pulling experience are unable to solve the broken string task (Taylor et al., 2012).

      However, when a green table was placed behind the string to obscure the “lollipop” structure during the training, the bees could not see the “lollipop” during the initial training stage or after pulling the string from under the table. In this situation, the bees were unable to identify the connected string, further proving that bumblebees chose the connected string based on image matching.

      Line 68: suggest remove 'meticulously'.

      We removed “meticulously”.

      Line 99: This is an exciting finding, can the authors please provide a video of a bee solving this task on its first trial?

      We added videos in the supplementary materials.

      Line 133: perceptual-motor feedback loops should be introduced in the introduction.

      We introduced perceptual-motor feedback loops in the revised manuscript.

      Line 136: please clarify the prior experience of these bees, it is not clear from the text.

      We clarified the prior experience of these bees as follow: Bumblebees were initially attracted to feed on yellow artificial flowers, and then trained with transparent tables covered by black tape (S7 video) through a four-step process.

      Line 138: from the video it is not possible to see the bee's perspective of this occlusion. Do the authors have a video or image showing the feedback the bees received? I think this is highly important if they wish to argue that this condition prevents the use of both image matching and a perceptual-motor feedback loop.

      We prevented the use of image matching: the bees were unable to see the flower moving towards them above the table during the training phase in this condition. But the bees may receive visual image both after pulling the string out from the table and in the initial stages of training in this condition.

      Line 147: please clarify what experience these bees had before this test.

      We added the prior experience of bumblebees before training as follow: We therefore designed further experiments based on Taylor et al. (2012) to test this hypothesis. Bumblebees were first trained to feed on yellow artificial, and then trained with the same procedure as Experiment 2, but the connected strings were coiled in the test.

      Line 155: This is a highly similar test to that used in Taylor et al 2012, have the authors seen this study?

      We mentioned the reference in the revised manuscript as follows: We therefore designed further experiments based on Taylor et al. (2012) to test this hypothesis.

      Line 183: This sentence needs rewriting "Since the vast majority of animals, including dogs 183 (Osthaus et al., 2005), cats (Whitt et al., 2009), western scrub-jays (Hofmann et al.,2016) and azure-winged magpies (Wang et al., 2019) are failing in such tasks spontaneously".

      We changed the sentence as suggested by the reviewer as follow:  Some animals, including dogs (Osthaus et al., 2005), cats (Whitt et al., 2009), western scrub-jays (Hofmann et al., 2016) and azure-winged magpies (Wang et al., 2019) fail in such task spontaneously.

      Line 186: "complete comprehension of the functionality of strings is rare" I am not sure the evidence in the current literature supports any animal showing full understanding, can the authors explain how they reach this conclusion?

      We wished to say that few animal species could distinguish between connected and disconnected strings without trial and error learning. We revised the sentence as follows:

      It is worth noting that some crows and parrots known for complex cognition perform poorly on broken string task without perceptual feedback or learning. For example, New Caledonian crows use perceptual feedback strategies to solve broken string-pulling task, and no individual showed a significant preference for the connected string when perceptual feedback is restricted (Taylor et al., 2012). Some Australian magpies and African grey parrots can solve the broken string task, but it required a high number of trials, indicating that learning plays a crucial role in solving this task (Molina et al., 2019; Johnsson et al., 2023).

      Line 190: the authors need to clarify which part of their study provides positive evidence for this conclusion.

      We added the evidence for this conclusion as follows: Our findings suggest that bumblebees with experience of string pulling prefer the connected strings, but they failed to identify the interrupted strings when the string was coiled in the test.

      Line 265: was the far end of the string glued only?

      The entire string was glued to the floor, not just the far ends of the string.

    1. "Don't Download This Song" is the first single from "Weird Al" Yankovic's 12th studio album Straight Outta Lynwood. The song was released exclusively on August 21, 2006 as a digital download. It is a style parody of "We Are the World", "Voices That Care", "Hands Across America", "Heal the World" and other similar charity songs. The song "describes the perils of online music file-sharing" in a tongue-in-cheek manner.[1] To further the sarcasm, the song was freely available for streaming and to legally download in DRM-free MPEG fileformat at Weird Al's Myspace page, a standalone website,[2] as well as his YouTube channel. Background[edit] "Don't Download This Song" references several court cases related to the RIAA and copyright infringement of music. Among these are lawsuits against "a grandma" (presumably Gertrude Walton,[3] who was sued for copyright infringement six months after dying) and a "7-year-old girl" (presumably a reference to Tanya Andersen's daughter[4] sued at age 10 for alleged copyright infringements made at the age of 7), as well as Lars Ulrich's strong stance against copyright infringement of music in the days of Napster. The song also challenges the RIAA's claim that file sharing prevents the artists from profiting from their work, as the song argues that they are still very financially successful via their recording contracts: ("Don't take away money from artists just like me/How else can I afford another solid-gold Humvee, And diamond-studded swimming pools? These things don't grow on trees"). Mention is also made of Tommy Chong's time spent in prison.[5] Yankovic's own views on filesharing are less clear-cut: .mw-parser-output .templatequote{overflow:hidden;margin:1em 0;padding:0 32px}.mw-parser-output .templatequote .templatequotecite{line-height:1.5em;text-align:left;padding-left:1.6em;margin-top:0}I have very mixed feelings about it. On one hand, I’m concerned that the rampant downloading of my copyright-protected material over the Internet is severely eating into my album sales and having a decidedly adverse effect on my career. On the other hand, I can get all the Metallica songs I want for FREE! WOW!!!!!— "Weird Al" Yankovic, "Ask Al" Q&As for May 2000 Yankovic's intention was to leave the listener with no clear understanding of Yankovic's own views on the matter, "all by design".[6] Music video[edit] The death scene from the music video. The music video, animated by Bill Plympton, premiered August 22, 2006 on Yahoo! Music. It depicts the vision of the capture, trial, imprisonment, attempted execution, escape, and burning of a young boy who burns a CD on his computer.[7] The boy's death, where he stands on top of a tower just before it explodes, parodies the film White Heat, where Cody Jarrett, played by James Cagney, dies in a similar fashion. Various people, from policemen to criminals to even sharks and dogs, are then seen celebrating throughout the ending chorus. But at the end, it turns out the boy is just imagining what would happen if he downloaded the song, so he throws away the burned CD and goes back to playing his guitar. Throughout the song, the video coloring gradually changes from color to grayscale to dark grayscale to yellowed. On MTV's MTV Music site where this music video is available, they have censored the names of the file sharing programs in the song, such as LimeWire or KaZaA.[8] Weird Al explained that MTV contacted him and told him they would not air his video if the references to the filesharing programs were not in some way removed, so he "made the creative decision to bleep them out as obnoxiously as possible, so that there would be no mistake I was being censored."[9] The video was praised by the Annie Awards and was subsequently nominated for Best Animated Short Subject for its 34th ceremony, but was beat out by the Ice Age featurette, No Time for Nuts. See also[edit] List of singles by "Weird Al" Yankovic List of songs by "Weird Al" Yankovic References[edit] .mw-parser-output .reflist{margin-bottom:0.5em;list-style-type:decimal}@media screen{.mw-parser-output .reflist{font-size:90%}}.mw-parser-output .reflist .references{font-size:100%;margin-bottom:0;list-style-type:inherit}.mw-parser-output .reflist-columns-2{column-width:30em}.mw-parser-output .reflist-columns-3{column-width:25em}.mw-parser-output .reflist-columns{margin-top:0.3em}.mw-parser-output .reflist-columns ol{margin-top:0}.mw-parser-output .reflist-columns li{page-break-inside:avoid;break-inside:avoid-column}.mw-parser-output .reflist-upper-alpha{list-style-type:upper-alpha}.mw-parser-output .reflist-upper-roman{list-style-type:upper-roman}.mw-parser-output .reflist-lower-alpha{list-style-type:lower-alpha}.mw-parser-output .reflist-lower-greek{list-style-type:lower-greek}.mw-parser-output .reflist-lower-roman{list-style-type:lower-roman} ^ Bill Plympton Studio Archived November 16, 2006, at the Wayback Machine ^ .mw-parser-output cite.citation{font-style:inherit;word-wrap:break-word}.mw-parser-output .citation q{quotes:"\"""\"""'""'"}.mw-parser-output .citation:target{background-color:rgba(0,127,255,0.133)}.mw-parser-output .id-lock-free.id-lock-free a{background:url("//upload.wikimedia.org/wikipedia/commons/6/65/Lock-green.svg")right 0.1em center/9px no-repeat}.mw-parser-output .id-lock-limited.id-lock-limited a,.mw-parser-output .id-lock-registration.id-lock-registration a{background:url("//upload.wikimedia.org/wikipedia/commons/d/d6/Lock-gray-alt-2.svg")right 0.1em center/9px no-repeat}.mw-parser-output .id-lock-subscription.id-lock-subscription a{background:url("//upload.wikimedia.org/wikipedia/commons/a/aa/Lock-red-alt-2.svg")right 0.1em center/9px no-repeat}.mw-parser-output .cs1-ws-icon a{background:url("//upload.wikimedia.org/wikipedia/commons/4/4c/Wikisource-logo.svg")right 0.1em center/12px no-repeat}body:not(.skin-timeless):not(.skin-minerva) .mw-parser-output .id-lock-free a,body:not(.skin-timeless):not(.skin-minerva) .mw-parser-output .id-lock-limited a,body:not(.skin-timeless):not(.skin-minerva) .mw-parser-output .id-lock-registration a,body:not(.skin-timeless):not(.skin-minerva) .mw-parser-output .id-lock-subscription a,body:not(.skin-timeless):not(.skin-minerva) .mw-parser-output .cs1-ws-icon a{background-size:contain;padding:0 1em 0 0}.mw-parser-output .cs1-code{color:inherit;background:inherit;border:none;padding:inherit}.mw-parser-output .cs1-hidden-error{display:none;color:var(--color-error,#d33)}.mw-parser-output .cs1-visible-error{color:var(--color-error,#d33)}.mw-parser-output .cs1-maint{display:none;color:#085;margin-left:0.3em}.mw-parser-output .cs1-kern-left{padding-left:0.2em}.mw-parser-output .cs1-kern-right{padding-right:0.2em}.mw-parser-output .citation .mw-selflink{font-weight:inherit}@media screen{.mw-parser-output .cs1-format{font-size:95%}html.skin-theme-clientpref-night .mw-parser-output .cs1-maint{color:#18911f}}@media screen and (prefers-color-scheme:dark){html.skin-theme-clientpref-os .mw-parser-output .cs1-maint{color:#18911f}}"Weird Al- Dont Download This Song". 2007-02-26. Archived from the original on 2007-02-26. Retrieved 2022-12-13. ^ "RIAA sues the dead". The Register. Retrieved 18 April 2018. ^ Beckerman, Ray (23 March 2007). "Recording Industry vs The People: RIAA Insists on Deposing Tanya Andersen's 10-year-old daughter". Retrieved 18 April 2018. ^ "kuro5hin.org". www.kuro5hin.org. Retrieved 18 April 2018. ^ Rabin, Nathan (2011-06-29). ""Weird Al" Yankovic". The A.V. Club. Retrieved 2011-06-29. ^ Premieres on Yahoo! Music Archived 2006-08-21 at the Wayback Machine ^ "MTV Bleeps File Sharing Software Out Of Music Videos". 30 October 2008. Retrieved 18 April 2018. ^ Cohen, Noam (2 November 2008). "Censorship, or What Really Weirds Out Weird Al". The New York Times. Retrieved 18 April 2018. External links[edit] alyankovicVEVO, "Weird Al Yankovic - Don't Download This Song", YouTube, October 2, 2009. The music video at Yankovic's official YouTube Vevo website. Plymptoons, DON'T DOWNLOAD THIS SONG - Weird Al Yankovic & Bill Plympton, YouTube. The music video at Bill Plympton's official YouTube website. Listen to the Song and Send E-Cards .mw-parser-output .navbox{box-sizing:border-box;border:1px solid #a2a9b1;width:100%;clear:both;font-size:88%;text-align:center;padding:1px;margin:1em auto 0}.mw-parser-output .navbox .navbox{margin-top:0}.mw-parser-output .navbox+.navbox,.mw-parser-output .navbox+.navbox-styles+.navbox{margin-top:-1px}.mw-parser-output .navbox-inner,.mw-parser-output .navbox-subgroup{width:100%}.mw-parser-output .navbox-group,.mw-parser-output .navbox-title,.mw-parser-output .navbox-abovebelow{padding:0.25em 1em;line-height:1.5em;text-align:center}.mw-parser-output .navbox-group{white-space:nowrap;text-align:right}.mw-parser-output .navbox,.mw-parser-output .navbox-subgroup{background-color:#fdfdfd}.mw-parser-output .navbox-list{line-height:1.5em;border-color:#fdfdfd}.mw-parser-output .navbox-list-with-group{text-align:left;border-left-width:2px;border-left-style:solid}.mw-parser-output tr+tr>.navbox-abovebelow,.mw-parser-output tr+tr>.navbox-group,.mw-parser-output tr+tr>.navbox-image,.mw-parser-output tr+tr>.navbox-list{border-top:2px solid #fdfdfd}.mw-parser-output .navbox-title{background-color:#ccf}.mw-parser-output .navbox-abovebelow,.mw-parser-output .navbox-group,.mw-parser-output .navbox-subgroup .navbox-title{background-color:#ddf}.mw-parser-output .navbox-subgroup .navbox-group,.mw-parser-output .navbox-subgroup .navbox-abovebelow{background-color:#e6e6ff}.mw-parser-output .navbox-even{background-color:#f7f7f7}.mw-parser-output .navbox-odd{background-color:transparent}.mw-parser-output .navbox .hlist td dl,.mw-parser-output .navbox .hlist td ol,.mw-parser-output .navbox .hlist td ul,.mw-parser-output .navbox td.hlist dl,.mw-parser-output .navbox td.hlist ol,.mw-parser-output .navbox td.hlist ul{padding:0.125em 0}.mw-parser-output .navbox .navbar{display:block;font-size:100%}.mw-parser-output .navbox-title .navbar{float:left;text-align:left;margin-right:0.5em}body.skin--responsive .mw-parser-output .navbox-image img{max-width:none!important}@media print{body.ns-0 .mw-parser-output .navbox{display:none!important}}show.mw-parser-output .navbar{display:inline;font-size:88%;font-weight:normal}.mw-parser-output .navbar-collapse{float:left;text-align:left}.mw-parser-output .navbar-boxtext{word-spacing:0}.mw-parser-output .navbar ul{display:inline-block;white-space:nowrap;line-height:inherit}.mw-parser-output .navbar-brackets::before{margin-right:-0.125em;content:"[ "}.mw-parser-output .navbar-brackets::after{margin-left:-0.125em;content:" ]"}.mw-parser-output .navbar li{word-spacing:-0.125em}.mw-parser-output .navbar a>span,.mw-parser-output .navbar a>abbr{text-decoration:inherit}.mw-parser-output .navbar-mini abbr{font-variant:small-caps;border-bottom:none;text-decoration:none;cursor:inherit}.mw-parser-output .navbar-ct-full{font-size:114%;margin:0 7em}.mw-parser-output .navbar-ct-mini{font-size:114%;margin:0 4em}html.skin-theme-clientpref-night .mw-parser-output .navbar li a abbr{color:var(--color-base)!important}@media(prefers-color-scheme:dark){html.skin-theme-clientpref-os .mw-parser-output .navbar li a abbr{color:var(--color-base)!important}}@media print{.mw-parser-output .navbar{display:none!important}}vte"Weird Al" Yankovic "Weird Al" Yankovic Jon "Bermuda" Schwartz Steve Jay Jim West Rubén Valtierra Rick Derringer Studio albums "Weird Al" Yankovic "Weird Al" Yankovic in 3-D Dare to Be Stupid Polka Party! Even Worse UHF – Original Motion Picture Soundtrack and Other Stuff Off the Deep End Alapalooza Bad Hair Day Running with Scissors Poodle Hat Straight Outta Lynwood Alpocalypse Mandatory Fun Soundtrack albums Weird: The Al Yankovic Story EPs Another One Rides the Bus Internet Leaks Compilations Greatest Hits The Best of Yankovic The Food Album Permanent Record: Al in the Box Greatest Hits Vol. II The TV Album The Essential "Weird Al" Yankovic Squeeze Box: The Complete Works of "Weird Al" Yankovic Songs "My Bologna" "Another One Rides the Bus" "Ricky" "I Love Rocky Road" "Eat It" "I Lost on Jeopardy" "Like a Surgeon" "Yoda" "Hooked on Polkas" "Dare to Be Stupid" "I Want a New Duck "Living with a Hernia" "Christmas at Ground Zero" "Fat" "Money for Nothing/Beverly Hillbillies" "Chicken Pot Pie" "Smells Like Nirvana" "You Don't Love Me Anymore" "Jurassic Park" "Bedrock Anthem" "Achy Breaky Song" "Headline News" "Amish Paradise" "Spy Hard" "The Night Santa Went Crazy" "The Saga Begins" "It's All About the Pentiums" "Polka Power!" "Pretty Fly for a Rabbi" "Albuquerque" "Bob" "Couch Potato" "eBay" "You're Pitiful" "Don't Download This Song" "White & Nerdy" "Pancreas" "Canadian Idiot" "Trapped in the Drive-Thru" "Whatever You Like" "Craigslist" "Perform This Way" "Tacky" "Word Crimes" "Foil" "Handy" "First World Problems" Videography Al TV The Compleat Al UHF The "Weird Al" Yankovic Video Library Alapalooza: The Videos Bad Hair Day: The Videos The Weird Al Show "Weird Al" Yankovic: The Ultimate Video Collection "Weird Al" Yankovic Live!: The Alpocalypse Tour Tours An Evening of Dementia with Dr. Demento in Person Plus "Weird Al" Yankovic Mandatory World Tour Ridiculously Self-Indulgent, Ill-Advised Vanity Tour Strings Attached Tour The Unfortunate Return of the Ridiculously Self-Indulgent, Ill-Advised Vanity Tour Related articles Discography Videography Polka medleys Peter & the Wolf/Carnival of the Animals – Part II Weird: The Al Yankovic Story Category showvteFilms directed by Bill PlymptonFeature films The Tune (1992) I Married a Strange Person! (1998) Mutant Aliens (2001) Hair High (2004) Idiots and Angels (2008) Cheatin' (2013) Hitler's Folly (2016) Revengeance (2016) Short films Your Face (1987) 12 Tiny Christmas Tales (2001) Guard Dog (2004) The Cow Who Wanted to Be a Hamburger (2010) Music videos "Heard 'Em Say" (2004) "Don't Download This Song" (2005) "TMZ" (2011) Authority control databases MusicBrainz release groupMusicBrainz work <img src="https://login.wikimedia.org/wiki/Special:CentralAutoLogin/start?type=1x1" alt="" width="1" height="1" style="border: none; position: absolute;"> Retrieved from "https://en.wikipedia.org/w/index.php?title=Don%27t_Download_This_Song&oldid=1239009310" Categories: 2006 singlesProtest songs"Weird Al" Yankovic songsSongs written by "Weird Al" YankovicPop balladsMusic videos directed by Bill PlymptonSongs about the Internet2006 songs2000s balladsAnimated music videosVolcano Entertainment singlesHidden categories: Webarchive template wayback linksArticles with short descriptionShort description matches WikidataArticles with hAudio microformatsArticles with MusicBrainz release group identifiersArticles with MusicBrainz work identifiers This page was last edited on 6 August 2

      you heard weird al.dont pirate or download his song!!!

    1. Rather than assuming that the entrepreneurialpersonality can be characterized by set of unified traits, Steyaert (2007a) contends thatthe question should be approached from a narrative point of view

      I really like this angle (and it's why I highlighted the "narrative turn" on pg. 230. This article engages in a "selective reading" (I love that term) of Branson's narrative (literally, his autobiography). But every account of entrepreneurs and their successes (and failures/struggles to succeed) can be understood as a story. We're not just reading the story of that individual (though that's often the focus); we are also encountering the story of entrepreneurship and the process of its ongoing fashioning/re-imagining. When I ask you to consider how you're entrepreneurial, I'm asking you to re-write your own story...

    2. create structures of desire that teach us how to desireto become an entrepreneur

      all of the success stories surrounding entrepreneurs encourage us to want to be entrepreneurs... and not just the heroic version either! All the narratives of people with side-hustles teach us to desire a similar pursuit, selling a gig-economy and precarious labour under the guise of passion and self-improvement... It's not just that we are all entrepreneurs (Szeman), but that we desire to be... (because we all DESIRE (or ought to desire, supposedly) more success, more freedom, more independence (financially and otherwise)...)

      And maybe, the more we desire and dream, the more we might do (converting our desires into action and making our dreams reality !!)

    3. the desire for transgression (overcoming oneself) and the desire forauthenticity (becoming oneself) make up the entrepreneurial subjectivity

      not just in Branson's case either. I think it's reasonable to suggest that these are potentially pillars in any entrepreneurial mindset.

    1. “cruel optimism”

      I think this is a very powerful concept -- as a former student said, entrepreneurs must have cruel optimism, individuals must be able to adapt and ride out whatever wave or obstacle comes around.

      Berlant uses the term cruel optimism to refer to our our investments in “compromised conditions of possibility whose realization is discovered to be impossible, sheer fantasy.” (i.e., we keep cheering for a team we know will lose; we maintain hope in an unattainable romantic ideal promulgated by Hollywood or pursue happiness based on unrealistic beauty standards; we engage in small acts of environmental stewardship like recycling or buying a hybrid in the face of potentially unstoppable climate change...) Berlant basically means that the thing we seek to achieve, the thing (or state of being) that we desire (or the act of seeking and desiring itself) might actually threaten our well-being (that's what makes it cruel!). As she put it succinctly, “a relation of cruel optimism exists when something you desire is actually an obstacle to your flourishing.”

      This relates to entrepreneurialism in so many ways: Engaging in the gig economy or a side-hustle as a way to increase one's income (or security) in uncertain times is cruel and optimistic. Similarly, we encounter aspirational labour in the form of internships or any form of unpaid labour while looking for a "real" job. Perhaps you feel the pressure of cultivating a sense of employability. According to Frayne (2015), today, students are expected “to improve their prospects by training, acquiring educational credentials, networking, learning how to project the right kind of personality, and gaining life experiences that match up with the values sought by employers.” In other words, they have to act entrepreneurially even to get a non-entrepreneurial job. As Hawzen et al. (2018) assert, this incites anxiety and results in a colonization of one’s entire life by work-related demands as students feel the need to separate themselves from the competition, doing things like volunteering to gain an advantage or to get a "foot in the door"... We also see it to a certain extent in the example of entrepreneurial vloggers in the sense that the fantasy of a “good life” through fame and fortune is rarely realized. The cruel conditions of precocity are, for most, more of a reality than the fantasy... and we take up this theme explicitly in two weeks hence with digital 'autopreneurs'

      Overall, this also highlights one of the reasons I chose this article -- rather than just highlighting how entrepreneurs are certain types of people (or motivated by certain types of things), it emphasizes how entrepreneurship is a mental orientation, not just a business concept but a way of living. But it's not all sunshine and happiness. Cruel optimism, indeed!

      What about you? Are you familiar with the feeling of 'cruel optimism'? Does it define the current times or your current disposition?

    2. The status of entrepreneurship as a new common sense of subjectivity and economic practice

      Remember at the beginning of the article (when Szeman says "we are all entrepreneurs now") (p. 472)? He doesn't mean that we are all creating business start-ups. Rather, he's suggesting that there is a spirit-of-the-times wherein entrepreneurship has become this new common-sense reality. It is both a dominant way of thinking about how we ought to act, AND an informal rulebook for how economies (and other forms of practice) ought to function too... In other words, entrepreneurship isn't just about undertaking profit-making (and risk-inducing) economic practices in capitalism. Rather, it's about undertaking a new subjectivity, a new identity when it comes to how we think of ourselves, how we relate to others, and how we respond to our wider social, cultural, political, and economic environment.

    1. Melissa Rich, a freshman at Stevens Institute of Technology, says she felt the effects of lowered expectations in her own high-school classes during the pandemic. “I used to be, in middle school at least, the kid that would always get stuff in on time and cared a lot about my classes,” says Rich, who was in ninth grade when Covid hit. “The pandemic changed the way I worked. It definitely stunted me a little bit.”A marine-biology class she took in high school consisted of worksheets. A geometry teacher would let students use a tutoring platform during tests to figure out answers. “I take responsibility for this. I can’t say it’s the teacher’s fault for not pushing me hard enough,” she says, “but when people would just let us do anything, I did not feel motivated to do extra work for classes I wasn’t passionate about.”

      I felt this way during the pandemic as well. There was a shift in the way classes were taught and in the quality of work students were expected to do. I always wanted to do my best, and I was always pushing myself. However, during the pandemic, it felt like teachers didn't care as much if you succeeded or not, and they didn't care if you actually learned.

    1. If they love learning, they will persist through challenges!

      Im not sure that it is as simple as this. There is more that goes into students ability to persist through challenges than just a love for learning. Sometimes we have students that love learning but are unable to persist through challenge because of embarrassment or what has been previously told to them about how they are as a student. As teachers we want to build up students identities as learners to help them persist. I don't think it's always as simple as the student loving what they are doing.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews: 

      Reviewer #1 (Public Review): 

      Summary: 

      Glaser et al present ExA-SPIM, a light-sheet microscope platform with large volumetric coverage (Field of view 85mm^2, working distance 35mm), designed to image expanded mouse brains in their entirety. The authors also present an expansion method optimized for whole mouse brains and an acquisition software suite. The microscope is employed in imaging an expanded mouse brain, the macaque motor cortex, and human brain slices of white matter. 

      This is impressive work and represents a leap over existing light-sheet microscopes. As an example, it offers a fivefold higher resolution than mesoSPIM (https://mesospim.org/), a popular platform for imaging large cleared samples. Thus while this work is rooted in optical engineering, it manifests a huge step forward and has the potential to become an important tool in the neurosciences. 

      Strengths: 

      - ExA-SPIM features an exceptional combination of field of view, working distance, resolution, and throughput. 

      - An expanded mouse brain can be acquired with only 15 tiles, lowering the burden on computational stitching. That the brain does not need to be mechanically sectioned is also seen as an important capability. 

      - The image data is compelling, and tracing of neurons has been performed. This demonstrates the potential of the microscope platform. 

      Weaknesses: 

      - There is a general question about the scaling laws of lenses, and expansion microscopy, which in my opinion remained unanswered: In the context of whole brain imaging, a larger expansion factor requires a microscope system with larger volumetric coverage, which in turn will have lower resolution (Figure 1B). So what is optimal? Could one alternatively image a cleared (non-expanded) brain with a high-resolution ASLM system (Chakraborty, Tonmoy, Nature Methods 2019, potentially upgraded with custom objectives) and get a similar effective resolution as the authors get with expansion? This is not meant to diminish the achievement, but it was unclear if the gains in resolution from the expansion factor are traded off by the scaling laws of current optical systems. 

      Paraphrasing the reviewer: Expanding the tissue requires imaging larger volumes and allows lower optical resolution. What has been gained?

      The answer to the reviewer’s question is nuanced and contains four parts. 

      First, optical engineering requirements are more forgiving for lenses with lower resolution. Lower resolution lenses can have much larger fields of view (in real terms: the number of resolvable elements, proportional to ‘etendue’) and much longer working distances. In other words, it is currently more feasible to engineer lower resolution lenses with larger volumetric coverage, even when accounting for the expansion factor. 

      Second, these lenses are also much better corrected compared to higher resolution (NA) lenses. They have a flat field of view, negligible pincushion distortions, and constant resolution across the field of view. We are not aware of comparable performance for high NA objectives, even when correcting for expansion.

      Third, although clearing and expansion render tissues ‘transparent’, there still exist refractive index inhomogeneities which deteriorate image quality, especially at larger imaging depths. These effects are more severe for higher optical resolutions (NA), because the rays entering the objective at higher angles have longer paths in the tissue and will see more aberrations. For lower NA systems, such as ExaSPIM, the differences in paths between the extreme and axial rays are relatively small and image formation is less sensitive to aberrations. 

      Fourth, aberrations are proportional to the index of refraction inhomogeneities (dn/dx). Since the index of refraction is roughly proportional to density, scattering and aberration of light decreases as M^3, where M is the expansion factor. In contrast, the imaging path length through the tissue only increases as M. This produces a huge win for imaging larger samples with lower resolutions. 

      To our knowledge there are no convincing demonstrations in the literature of diffraction-limited ASLM imaging at a depth of 1 cm in cleared mouse brain tissue, which would be equivalent to the ExA-SPIM imaging results presented in this manuscript.  

      In the discussion of the revised manuscript we discuss these factors in more depth. 

      - It was unclear if 300 nm lateral and 800 nm axial resolution is enough for many questions in neuroscience. Segmenting spines, distinguishing pre- and postsynaptic densities, or tracing densely labeled neurons might be challenging. A discussion about the necessary resolution levels in neuroscience would be appreciated. 

      We have previously shown good results in tracing the thinnest (100 nm thick) axons over cm scales with 1.5 um axial resolution. It is the contrast (SNR) that matters, and the ExaSPIM contrast exceeds the block-face 2-photon contrast, not to mention imaging speed (> 10x).  

      Indeed, for some questions, like distinguishing fluorescence in pre- and postsynaptic structures, higher resolutions will be required (0.2 um isotropic; Rah et al Frontiers Neurosci, 2013). This could be achieved with higher expansion factors.

      This is not within the intended scope of the current manuscript. As mentioned in the discussion section, we are working towards ExA-SPIM-based concepts to achieve better resolution through the design and fabrication of a customized imaging lens that maintains a high volumetric coverage with increased numerical aperture.  

      - Would it be possible to characterize the aberrations that might be still present after whole brain expansion? One approach could be to image small fluorescent nanospheres behind the expanded brain and recover the pupil function via phase retrieval. But even full width half maximum (FWHM) measurements of the nanospheres' images would give some idea of the magnitude of the aberrations. 

      We now included a supplementary figure highlighting images of small axon segments within distal regions of the brain.  

      Reviewer #2 (Public Review)

      Summary: 

      In this manuscript, Glaser et al. describe a new selective plane illumination microscope designed to image a large field of view that is optimized for expanded and cleared tissue samples. For the most part, the microscope design follows a standard formula that is common among many systems (e.g. Keller PJ et al Science 2008, Pitrone PG et al. Nature Methods 2013, Dean KM et al. Biophys J 2015, and Voigt FF et al. Nature Methods 2019). The primary conceptual and technical novelty is to use a detection objective from the metrology industry that has a large field of view and a large area camera. The authors characterize the system resolution, field curvature, and chromatic focal shift by measuring fluorescent beads in a hydrogel and then show example images of expanded samples from mouse, macaque, and human brain tissue. 

      Strengths: 

      I commend the authors for making all of the documentation, models, and acquisition software openly accessible and believe that this will help assist others who would like to replicate the instrument. I anticipate that the protocols for imaging large expanded tissues (such as an entire mouse brain) will also be useful to the community. 

      Weaknesses: 

      The characterization of the instrument needs to be improved to validate the claims. If the manuscript claims that the instrument allows for robust automated neuronal tracing, then this should be included in the data. 

      The reviewer raises a valid concern. Our assertion that the resolution and contrast is sufficient for robust automated neuronal tracing is overstated based on the data in the paper. We are hard at work on automated tracing of datasets from the ExA-SPIM microscope. We have demonstrated full reconstruction of axonal arbors encompassing >20 cm of axonal length.  But including these methods and results is out of the scope of the current manuscript. 

      The claims of robust automated neuronal tracing have been appropriately modified.  

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors): 

      Smaller questions to the authors: 

      - Would a multi-directional illumination and detection architecture help? Was there a particular reason the authors did not go that route?

      Despite the clarity of the expanded tissue, and the lower numerical aperture of the ExA-SPIM microscope, image quality still degrades slightly towards the distal regions of the brain relative to both the excitation and detection objective. Therefore, multi-directional illumination and detection would be advantageous. Since the initial submission of the manuscript, we have undertaken re-designing the optics and mechanics of the system. This includes provisions for multi-directional illumination and detection. However, this new design is beyond the scope of this manuscript. We now mention this in L254-255 of the Discussion section.

      - Why did the authors not use the same objective for illumination and detection, which would allow isotropic resolution in ASLM? 

      The current implementation of ASLM requires an infinity corrected objective (i.e. conjugating the axial sweeping mechanism to the back focal plane). This is not possible due to the finite conjugate design of the ExA-SPIM detection lens.

      More fundamentally, pushing the excitation NA higher would result in a shorter light sheet Rayleigh length, which would require a smaller detection slit (shorter exposure time, lower signal to noise ratio). For our purposes an excitation NA of 0.1 is an excellent compromise between axial resolution, signal to noise ratio, and imaging speed. 

      For other potentially brighter biological structures, it may be possible to design a custom infinity corrected objective that enables ASLM with NA > 0.1.

      - Have the authors made any attempt to characterize distortions of the brain tissue that can occur due to expansion? 

      We have not systematically characterized the distortions of the brain tissue pre and post expansion. Imaged mouse brain volumes are registered to the Allen CCF regardless of whether or not the tissue was expanded. It is beyond the scope of this manuscript to include these results and processing methods, but we have confirmed that the ExA-SPIM mouse brain volumes contain only modest deformation that is easily accounted for during registration to the Allen CCF. 

      - The authors state that a custom lens with NA 0.5-0.6 lens can be designed, featuring similar specifications. Is there a practical design? Wouldn't such a lens be more prone to Field curvature? 

      This custom lens has already been designed and is currently being fabricated. The lens maintains a similar space bandwidth product as the current lens (increased numerical aperture but over a proportionally smaller field of view). Over the designed field of view, field curvature is <1 µm. However, including additional discussion or results of this customized lens is beyond the scope of this manuscript.

      Reviewer #2 (Recommendations For The Authors): 

      • System characterization: 

      - Please state what wavelength was used for the resolution measurements in Figure 2.

      An excitation wavelength of 561 nm was used. This has been added to the manuscript text.

      - The manuscript highlights that a key advance for the microscope is the ability to image over a very large 13 mm diameter field of view. Can the authors clarify why they chose to characterize resolution over an 8diameter mm field rather than the full area? 

      The 13 mm diameter field of view refers to the diagonal of the 10.6 x 8.0 mm field of view. The results presented in Figure 1c are with respect to the horizontal x direction and vertical y direction. A note indicating that the 13 mm is with respect to the diagonal of the rectangular imaging field has been added to the manuscript text. The results were presented in this way to present the axial and lateral resolution as a function of y (the axial sweeping direction).

      - The resolution estimates seem lower than I would expect for a 0.30 NA lens (which should be closer to ~850 nm for 515 nm emission). Could the authors clarify the discrepancy? Is this predicted by the Zemax model and due to using the lens in immersion media, related to sampling size on the camera, or something else? It would be helpful if the authors could overlay the expected diffraction-limited performance together with the plots in Figure 2C. 

      As mentioned previously, the resolution measurements were performed with 561 nm excitation and an emission bandpass of ~573 – 616 nm (595 nm average). Based on this we would expect the full width half maximum resolution to be ~975 nm. The resolution is in fact limited by sampling on the camera. The 3.76 µm pixel size, combined with the 5.0X magnification results in a sampling of 752 nm. Based on the Nyquist the resolution is limited to ~1.5 µm. We have added clarifying statements to the text.

      - I'm confused about the characterization of light sheet thickness and how it relates to the measured detection field curvature. The authors state that they "deliver a light sheet with NA = 0.10 which has a width of 12.5 mm (FWHM)." If we estimate that light fills the 0.10 NA, it should have a beam waist (2wo) of ~3 microns (assuming Gaussian beam approximations). Although field curvature is described as "minimal" in the text, it is still ~10-15 microns at the edge of the field for the emission bands for GFP and RFP proteins. Given that this is 5X larger than the light sheet thickness, how do the authors deal with this? 

      The generated light sheet is flat, with a thickness of ~ 3 µm. This flat light sheet will be captured in focus over the depth of focus of the detection objective. The stated field curvature is within 2.5X the depth of focus of the detection lens, which is equivalent to the “Plan” specification of standard microscope objectives.

      - In Figure 2E, it would be helpful if the authors could list the exposure times as well as the total voxels/second for the two-camera comparison. It's also worth noting that the Sony chip used in the VP151MX camera was released last year whereas the Orca Flash V3 chosen for comparison is over a decade old now. I'm confused as to why the authors chose this camera for comparison when they appear to have a more recent Orca BT-Fusion that they show in a picture in the supplement (indicated as Figure S2 in the text, but I believe this is a typo and should be Figure S3). 

      This is a useful addition, and we have added exposure times to the plot. We have also added a note that the Orca Flash V3 is an older generation sCMOS camera and that newer variants exist. Including the Orca BT-Fusion. The BT-Fusion has a read noise of 1.0 e- rms versus 1.6 e- rms, and a peak quantum efficiency of ~95% vs. 85%. Based on the discussion in Supplementary Note S1, we do not expect that these differences in specifications would dramatically change the data presented in the plot. In addition, the typo in Figure S2 has been corrected to Figure S3.

      - In Table S1, the authors note that they only compare their work to prior modalities that are capable of providing <= 1 micron resolution. I'm a bit confused by this choice given that Figure 2 seems to show the resolution of ExA-SPIM as ~1.5 microns at 4 mm off center (1/2 their stated radial field of view). It also excludes a comparison with the mesoSPIM project which at least to me seems to be the most relevant prior to this manuscript. This system is designed for imaging large cleared tissues like the ones shown here. While the original publication in 2019 had a substantially lower lateral resolution, a newer variant, Nikita et al bioRxiv (which is cited in general terms in this manuscript, but not explicitly discussed) also provides 1.5-micron lateral resolution over a comparable field of view. 

      We have updated the table to include the benchtop mesoSPIM from Nikita et al., Nature Communications, 2024. Based on this published version of the manuscript, the lateral resolution is 1.5 µm and axial resolution is 3.3 µm. Assuming the Iris 15 camera sensor, with the stated 2.5 fps, the volumetric rate (megavoxels/sec) is 37.41.

      - The authors state that, "We systematically evaluated dehydration agents, including methanol, ethanol, and tetrahydrofuran (THF), followed by delipidation with commonly used protocols on 1 mm thick brain slices. Slices were expanded and examined for clarity under a macroscope." It would be useful to include some data from this evaluation in the manuscript to make it clear how the authors arrived at their final protocol. 

      Additional details on the expansion protocol may be included in another manuscript.

      General comments: 

      • There is a tendency in the manuscript to use negative qualitative terms when describing prior work and positive qualitative terms when describing the work here. Examples include: 

      - "Throughput is limited in part by cumbersome and error-prone microscopy methods". While I agree that performing single neuron reconstructions at a large scale is a difficult challenge, the terms cumbersome and error-prone are qualitative and lacking objective metrics.

      We have revised this statement to be more precise, stating that throughput is limited in part by the speed and image quality of existing microscopy methods.

      - The resolution of the system is described in several places as "near-isotropic" whereas prior methods were described as "highly anisotropic". I agree that the ~1:3 lateral to axial ratio here is more isotropic than the 1:6 ratio of the other cited publications. However, I'm not sure I'd consider 3-fold worse axial resolution than lateral to be considered "near" isotropic.

      We agree that the term near-isotropic is ambiguous. We have modified the text accordingly, removing the term near-isotropic and where appropriate stating that the resolution is more isotropic than that of other cited publications.

      - exposures (which in the caption is described as "modest"). I'd suggest removing these qualitative terms and just stating the values.

      We agree and have changed the text accordingly.

      • The results section for Figure 5 is titled "Tracing axons in human neocortex and white matter". Although this section states "larger axons (>1 um) are well separated... allowing for robust automated and manual tracing" there is no data for any tracing in the manuscript. Although I agree that the images are visually impressive, I'm not sure that this claim is backed by data.

      We have now removed the text in this section referring to automated and manual tracing.

    1. To give just a sense of how different the emerging picture is: it is clear now that human societies before the advent of farming were not confined to small, egalitarian bands. On the contrary, the world of hunter-gatherers as it existed before the coming of agriculture was one of bold social experiments, resembling a carnival parade of polit- ical forms, far more than it does the drab abstractions of evolutionary theory. Agriculture, in turn, did not mean the inception of private property, nor did it mark an irreversible step towards inequality. In. fact, many of the first farming communities were relatively free of ranks and hierarchies. And far from setting class differences in stone, a surprising number of the world’s earliest cities were organized on robustly egalitarian lines, with no need for authoritarian rulers, ambi- tious warrior-politicians, or even bossy administrators. Information bearing on such issues has been pouring in from every quarter of the globe. As a result, researchers around the world have also been examining ethnographic and historical material in a new light. The pieces now exist to create an entirely different world history — but so far, they remain hidden to all but a few privileged experts (and even the experts tend to hesitate before abandoning their own tiny part of the puzzle, to compare notes with others out- side their specific subfield). Our aim in this book is to start putting some of the pieces of the puzzle together, in full awareness that nobody yet has anything like a complete set. The task is immense, and the issues so important, that it will take years of research and debate even to begin to understand the real implications of the pic- ture we're starting to see. But it’s crucial that we set the process in motion. One thing that will quickly become clear is that the preva- lent ‘big picture’ of history — shared by modern-day followers of Hobbes and Rousseau alike — has almost nothing to do with the g

      Some of the world’s first cities were organized without the need for rulers or hierarchies, challenging the idea that urbanization leads to inequality.

    1. But now we have more shows, so we’re still recapping a lot — it’s just that there’s so many more.

      It must be hard to pick something to recap considering how many shows people watch now.

    1. “cruel optimism”

      I think this is a very powerful concept -- as a former student said, entrepreneurs must have cruel optimism, individuals must be able to adapt and ride out whatever wave or obstacle comes around.

      Berlant uses the term cruel optimism to refer to our our investments in “compromised conditions of possibility whose realization is discovered to be impossible, sheer fantasy.” (i.e., we keep cheering for a team we know will lose; we maintain hope in an unattainable romantic ideal promulgated by Hollywood or pursue happiness based on unrealistic beauty standards; we engage in small acts of environmental stewardship like recycling or buying a hybrid in the face of potentially unstoppable climate change...) Berlant basically means that the thing we seek to achieve, the thing (or state of being) that we desire (or the act of seeking and desiring itself) might actually threaten our well-being (that's what makes it cruel!). As she put it succinctly, “a relation of cruel optimism exists when something you desire is actually an obstacle to your flourishing.”

      This relates to entrepreneurialism in so many ways: Engaging in the gig economy or a side-hustle as a way to increase one's income (or security) in uncertain times is cruel and optimistic. Similarly, we encounter aspirational labour in the form of internships or any form of unpaid labour while looking for a "real" job. Perhaps you feel the pressure of cultivating a sense of employability. According to Frayne (2015), today, students are expected “to improve their prospects by training, acquiring educational credentials, networking, learning how to project the right kind of personality, and gaining life experiences that match up with the values sought by employers.” In other words, they have to act entrepreneurially even to get a non-entrepreneurial job. As Hawzen et al. (2018) assert, this incites anxiety and results in a colonization of one’s entire life by work-related demands as students feel the need to separate themselves from the competition, doing things like volunteering to gain an advantage or to get a "foot in the door"... We also see it to a certain extent in the example of entrepreneurial vloggers in the sense that the fantasy of a “good life” through fame and fortune is rarely realized. The cruel conditions of precocity are, for most, more of a reality than the fantasy... and we take up this theme explicitly in two weeks hence with digital 'autopreneurs'

      Overall, this also highlights one of the reasons I chose this article -- rather than just highlighting how entrepreneurs are certain types of people (or motivated by certain types of things), it emphasizes how entrepreneurship is a mental orientation, not just a business concept but a way of living. But it's not all sunshine and happiness. Cruel optimism, indeed!

      What about you? Are you familiar with the feeling of 'cruel optimism'? Does it define the current times or your current disposition?

    2. The status of entrepreneurship as a new common sense of subjectivity and economic practice

      Remember at the beginning of the article (when Szeman says "we are all entrepreneurs now") (p. 472)? He doesn't mean that we are all creating business start-ups. Rather, he's suggesting that there is a spirit-of-the-times wherein entrepreneurship has become this new common-sense reality. It is both a dominant way of thinking about how we ought to act, AND an informal rulebook for how economies (and other forms of practice) ought to function too... In other words, entrepreneurship isn't just about undertaking profit-making (and risk-inducing) economic practices in capitalism. Rather, it's about undertaking a new subjectivity, a new identity when it comes to how we think of ourselves, how we relate to others, and how we respond to our wider social, cultural, political, and economic environment.

    1. Welcome back, this is part two of this lesson.

      We're going to continue immediately from the end of part one.

      So let's get started.

      Now let's talk about the first of the Glacier Storage Classes, S3 Glacier Instant Retrieval.

      If I had to summarize this storage class, it's like S3 standard in frequent access, except it offers cheaper storage, more expensive retrieval costs, and longer minimums.

      Standard IA is designed for when you need data instantly, but not very often, say once a month.

      Glacier Instant Retrieval extends this, so data where you still want instant retrieval, but where you might only access it say once every quarter.

      In line with this, it has a minimum storage duration charge of 90 days versus the 30 days of standard in frequent access.

      This class is the next step along the path of access frequency, as the access frequency of objects decrease, you can move them gradually from standard, then to standard in frequent access, and then to Glacier Instant Retrieval.

      The important thing to remember about this specific S3 Glacier class is that you still have instant access to your data.

      There's no retrieval process required, you can still use it like S3 standard and S3 standard in frequent access.

      It's just that it costs you more if you need to access the data, but less if you don't.

      Now let's move on to the next type of S3 Glacier Storage Class.

      And the next one I want to talk about is S3 Glacier Flexible Retrieval, and this storage class was formally known as S3 Glacier.

      The name was changed when the previously discussed Instant Retrieval class was added to the lineup of storage classes available within S3.

      So Glacier Flexible Retrieval has the same three availability zone architecture as S3 standard and S3 standard in frequent access.

      It has the same durability characteristics of 11-9s, and at the time of creating this lesson, S3 Glacier Flexible Retrieval has a storage cost which is about one-sixth of the cost of S3 standard.

      So it's really cost effective, but there are some serious trade-offs which you have to accept in order to make use of it.

      For the exam, it's these trade-offs which you need to be fully aware of.

      Conceptually, I want you to think of objects stored with the Glacier Flexible Retrieval class as cold objects.

      They aren't warm, they aren't ready for use, and this will form a good knowledge anchor for the exam.

      Now because they're cold, they aren't immediately available, they can't be made public.

      Well, you can see these objects within an S3 bucket, they're now just a pointer to that object.

      To get access to them, you need to perform a retrieval process.

      That's a specific operation, a job which needs to be run to gain access to the objects.

      Now you pay for this retrieval process.

      When you retrieve objects from S3 Glacier Flexible Retrieval, they're stored in the S3 standard in frequent access storage class on a temporary basis.

      You access them and then they're removed.

      You can retrieve them permanently by changing the class back to one of the S3 ones, but this is a different process.

      Now retrieval jobs come in three different types.

      We have expedited, which generally results in data being available within one to five minutes, and this is the most expensive.

      We've got standard where data is usually accessible in three to five hours, and then a low cost bulk option where data is available in between five and 12 hours.

      So the faster the job type, the more expensive.

      Now this means that S3 Glacier Flexible Retrieval has a first byte latency of minutes or hours, and that's really important to know for the exam.

      So while it's really cheap, you have to be able to tolerate, you can't make the objects public anymore, either in the bucket or using static website hosting, and two, when you do access the objects, it's not an immediate process.

      So you can see the object metadata in the bucket, but the data itself is in chilled storage, and you need to retrieve that data in order to access it.

      Now S3 Glacier Flexible Retrieval has some other limits, so a 40 kb minimum available size and a 90 day minimum available duration.

      For the exam, Glacier Flexible Retrieval is for situations where you need to store archival data where frequent or real-time access isn't needed.

      For example, yearly access, and you're OK with minutes to hours for retrieval operations.

      So it's one of the cheapest forms of storage in S3, as long as you can tolerate the characteristics of the storage class, but it's not the cheapest form of storage.

      That honor goes to S3 Glacier Deep Archive.

      Now S3 Glacier Deep Archive is much cheaper than the storage class we were just discussing.

      In exchange for that, there are even more restrictions which you need to be able to tolerate.

      Conceptually, where S3 Glacier Flexible Retrieval, which data in a chilled state, Glacier Deep Archive is data in a frozen state.

      Objects have minimum, so 40 kb minimum available size and 180 day minimum available duration.

      Like Glacier Flexible Retrieval, objects cannot be made publicly accessible.

      Access to the data requires a retrieval job.

      Just like Glacier Flexible Retrieval, the jobs temporarily restore to S3 standard and frequent access, but those retrieval jobs take longer.

      Standard is 12 hours and bulk is up to 48 hours, so this is much longer than Glacier Flexible Retrieval, and that's the compromise that you agree to.

      The storage is a lot cheaper in exchange for much longer restore times.

      Glacier Deep Archive should be used for data which is archival, which rarely, if ever, needs to be accessed, and where hours or days is tolerable for the retrieval process.

      So it's not really suited to primary system backups because of this restore time.

      It's more suited for secondary long-term archival backups or data which comes under legal or regulatory requirements in terms of retention length.

      Now this being said, there's one final type of storage class which I want to cover, and that's intelligent tearing.

      Now intelligent tearing is different from all the other storage classes which I've talked about.

      It's actually the storage class which contains five different storage tiers.

      With intelligent tearing, when you move objects into this class, there are a range of ways that an object can be stored.

      It can be stored within a frequent access tier or an infrequent access tier, or for objects which are accessed even less frequently, there's an archive instant access, archive access, or deep archived set of tiers.

      You can think of the frequent access tier like S3 standard and the infrequent access tier like S3 standard infrequent access, and the archive tiers are the same price of performance as S3, Glacier, instant retrieval, and flexible retrieval.

      And the deep archive tier is the same price of performance as Glacier Deep Archive.

      Now unlike the other S3 storage classes, you don't have to worry about moving objects between tiers.

      With intelligent tearing, the intelligent tearing system does this for you.

      Let's say that we have an object, say a picture of whiskers which is initially kind of popular and then not popular, and then it goes super viral.

      Well if you store this object using the intelligent tearing storage class, it would monitor the usage of the object.

      When the object is in regular use, it would stay within the frequent access tier and would have the same costs as S3 standard.

      If the object isn't accessed for 30 days, then it would be moved automatically into the infrequent tier where it would stay while being stored at a lower rate.

      Now at this stage you could also add configuration, so based on a bucket, prefix or object tag, any objects which are accessed less frequently can be moved into the three archive tiers.

      Now there's a 90 day minimum for archive instant access, and this is fully automatic.

      Think of this as a cheaper version of infrequent access for objects which are accessed even less frequently.

      Crucially this tier, so archive instant access, still gives you access to the data automatically as and when you need it, just like infrequent access.

      In addition to this, there are two more entirely optional tiers, archive access and deep archive.

      And these can be configured so that objects move into them when they haven't been accessed for 98 to 270 days for archive access, or 180 through to 730 days for deep archive.

      Now these are entirely optional, and it's worth mentioning that when objects are moved into these tiers, getting them back isn't immediate.

      There's a retrieval time to bring them back, so only use these tiers when your application can tolerate asynchronous access patterns.

      So archive instant access requires no application or system changes, it's just another tier for less frequently accessed objects with a lower cost.

      Archive access and deep archive changes things, your applications must support these tiers because retrieving objects requires specific API calls.

      Now if objects do stay in infrequent access or archive instant access, when the objects become super viral in access, these will be moved back to frequent access automatically with no retrieval charges.

      Intelligent tiering has a monitoring and automation cost per 1000 objects instead of the retrieval cost.

      So essentially the system manages the movement of data between these tiers automatically without any penalty for this management fee.

      The cost of the tiers are the same as the base S3 tiers, standard and infrequent access, there's just the management fee on top.

      So it's more flexible than S3 standard and S3 infrequent access, but it's more expensive because of the management fee.

      Now intelligent tiering is designed for long-lived data where the usage is...

      [Sounds of S3 storage] Changing or unknown, if the usage is static either frequently accessed or infrequently accessed, then you're better using the direct S3 storage class, either standard or infrequent access.

      Intelligent tiering is only good if you have data where the pattern changes or you don't know it.

      Now with that being said, that's all of the S3 storage classes which I want to cover.

      That's at least enough technical information and context which you'll need for the exam and to get started in the real world.

      So go ahead and complete the video and when you're ready, I look forward to you joining me in the next.

    1. Welcome back and in this demo lesson I just want to give you the opportunity to gain some practical experience of how S3 handles encryption.

      So what we're going to do is create an S3 bucket and into that bucket we're going to put a number of objects and for each one we're going to utilize a different type of server-side encryption.

      So we'll be uploading one object using SSE-S3 so S3 managed encryption and then one object using SSE-KMS which will utilize KMS for key management.

      So once we've uploaded those objects we'll experiment with some permissions changes just to see how each of these different encryption types work.

      So let's get started.

      Now the first thing you'll need to check is that you're logged into the IAM admin user of the general AWS account and you need to have the Northern Virginia region selected.

      Then let's move to S3 so click in the services drop-down type S3 and then open that in a new tab and then click to move to the S3 console.

      Now I want you to go ahead and click on create bucket and then just create a bucket called catpicks and then put some random text on the end.

      You should use something different and something that's unique to you.

      Leave the region set as US East Northern Virginia scroll all the way down to the bottom and click on create bucket.

      Next change across to the key management service console that will either be in your history or you'll have to type it in the services drop-down.

      Once you hear go ahead and click on create key, pick symmetric key then expand advanced options, make sure KMS is selected as well as single region key.

      Click next.

      The alias we'll be using is catpicks, so type catpicks and click next.

      Don't select anything for define key administrative permissions.

      We'll not set any permissions on the key policy this time so just click next and then on this screen define key usage permissions just click next again without selecting anything.

      So the key policy that's going to be created for this KMS key only trusts the account so only trust the account user of this specific AWS account and that's what we want.

      So go ahead and click on finish.

      At this point move back across to the S3 console, go into the catpicks bucket that you just created and I'll let you to go ahead and download the file that's attached to this lesson and then extract that file and inside the resultant folder are some objects that you're going to be uploading to this S3 bucket.

      So go ahead and do that and then click on upload and files then locate the folder that you just extracted and go into it and you should see three images in that folder default -merlin.jpg, sse -kms -ginny.jpg and sse -s3 -dwees.jpg.

      So that's good so the first one that we want to upload need to do these one by one because we're going to be configuring the encryption type to use.

      So the first one is sse -s3 -dwees.jpg so select that and click on open, expand properties and then scroll down to serve aside encryption and it's here where you can specify to accept the bucket defaults by not specifying an encryption key or you can specify an encryption key.

      Now when you pick to specify an encryption key you're again offered the ability to use the bucket default settings for encryption or you can override bucket settings and choose between two different types either Amazon s3 key which is sse -s3 or AWS key management service which is sse -kms.

      Now for this upload we're going to use sse -s3 so Amazon s3 managed keys so select this option scroll all the way down to the bottom and click upload.

      Wait for this upload process to complete and then click on close.

      Now we're going to follow that same process again so click on upload again, add files.

      This time sse -kms -ginny.

      So select that click on open, expand properties and then scroll down to serve aside encryption then click specify an encryption key override bucket settings for default encryption and this time we're going to use sse -kms so select that and then select choose from your AWS kms keys and then you can either use the AWS managed key so this is the service default key so AWS forward slash s3 or you can choose your own kms key to use for this object.

      What I'm going to do first is select this AWS managed key so the default key for this service and scroll all the way down to the bottom and click on upload.

      Wait for this upload process to complete and then click on close so that will encrypt the object using this default s3 AWS managed kms key that's now inside kms.

      I wanted to do that just to demonstrate how it automatically creates it so now let's go ahead and re-upload this object so click on upload, add files, select sse -kms -ginny.jpg and click on open, scroll down, expand properties, scroll down again, for service side encryption select specify an encryption key, select override bucket settings for default encryption, pick AWS key management service key so sse -kms, select choose from your AWS kms keys, click in the drop down and then select captics that we created earlier.

      Once you've done that scroll all the way down to the bottom click on upload, wait for that process to complete and then click on close.

      Now at this point we've got two different objects in this bucket and we're going to open both of these.

      We're going to start with sse -s3 -dwees so let's click on it and then click on open and that works and then let's try the other object sse -kms -ginny so click on that and click on open and that also opens okay because IAM admin is a full administrator of the entire AWS account so that includes s3 and all the services including KMS.

      Next what we're going to do is apply a deny policy on the IAM admin user which prevents us from using KMS so we'll stay as being a full account administrator and a full s3 administrator but we're going to block off the KMS service entirely and I want to demonstrate exactly what that does to our ability to open these three objects.

      So click on services and either open the IAM console from the history or type it in the find services box and then click it.

      Once we're here click on users select your IAM admin user, click add permissions and then create inline policy, click on the JSON tab and then delete the skeleton template that's in this box and then paste in the contents of the deny kms.json file and this is contained in the folder you extracted from this lessons download link and it's also attached to this lesson.

      This is what it should look like the effect is to deny it denies any actions KMS call on star so any KMS actions on all resources so essentially this blocks off the entire KMS service for this user.

      So go ahead and click on review policy call it deny KMS and click on create policy and this now means that IAM admin can no longer access KMS so now if we go back to the s3 console go inside the capex bucket we should still be able to open sse-s3 dweez.jpeg object.

      If we click that click on open because this is encrypted using sse-s3 and this is completely internal to the s3 product we should have no problems owning this object because we have the permissions inside s3 to do anything in s3 but something different happens if we try to open the sse-kms-ginning object.

      Now just to explain what will occur when I click on this open link s3 will then have to liaise with KMS and get KMS to decrypt the data encryption key that encrypts this object so we need to retrieve the encrypted data encryption key for this object and request that KMS decrypts it.

      Now if that worked we'd get back the plain text version of that key and we would use it to decrypt this object and it would open up in a tab without any issues because we've got full rights over s3 we have permission to do almost all of that process but what we don't have permission to do is to actually get KMS to decrypt this encrypted data encryption key.

      We don't have that permission because we just added a deny policy to the IAM admin user and as we know by now deny allow deny deny always wins and explicit deny always overrules everything else.

      So when I click on open and s3 retrieves this encrypted data encryption key and gives it to KMS and says please decrypt this and give me the plain text back KMS is going to refuse and what we see when we do that is we get an access deny.

      So now we've implemented this role separation so even though we have full s3 admin rights so if I went back to this bucket and I clicked on the sse-kms-ginning file and deleted it I would be able to delete that object because I have full control over s3 but I can't open it because I've prevented us accessing KMS and that's how we implement role separation.

      So sse-kms is definitely the encryption type that you want to use if you've got these really extensive regulations or any security requirements around key control.

      So let's go ahead and just remove that restriction so just go back to the IAM console.

      I just want to do this before we forget and have problems later in the course.

      Click on users click on IAM admin check the box next to deny KMS and then click remove and confirm that removal and that will allow us to access KMS again.

      We can verify that by moving to the KMS console and we can bring up this list which proves that we've got some access to KMS again so that's good.

      Now if we just go ahead and click on the AWS managed keys option on the left here this is where you'll be able to see this default encryption key that's used when you upload an object using sse-kms-incription but don't pick a particular key so this is now the default.

      Now if we open this because it's an AWS managed key we don't have the ability to set any key rotation we can see the key policy here but we can't make any changes to it this is set by AWS when it creates it so that it only allows accesses from s3 so this is a fixed key policy but we can't control anything about this key.

      Now contrast that with the customer managed keys that we've got and if we go into cat pics this is the key that we created now we can edit the key policy we could switch to policy view and make changes we've got the ability to control the key rotation so if you face any exam questions where you need to fully manage the keys that are used as part of the s3 encryption process then you've got to use sse-kms.

      Now if we just return to the s3 console there's just one more thing that I want to demonstrate go into the cat pics bucket again click on properties locate default encryption and then click on edit and this is where you get the option to specify the default encryption to use for this bucket.

      Now again this isn't a restriction this does not prevent anyone uploading objects to the bucket using a different type of encryption all it does is specify what the default is if the upload itself does not specify an encryption method so we could select Amazon s3 key which is sse-s3 and you might also see this referred to elsewhere as AES 256 it's also known by that name but we could also select the AWS key management service key known as sse-kms and this is where we can either choose to use the default key or pick a customer managed key that we want to use as the default for the bucket.

      So let's just demonstrate that go ahead and select the cat pics key to use for this bucket then scroll down and click on save changes and that will set the defaults for the bucket and we can demonstrate that let's go ahead and click on the objects tab and we're going to upload a new object to this bucket so click on upload add files then go ahead and select the default hyphen Merlin object and click open scroll down click on upload and even though we didn't pick a particular encryption method for this object it's going to use the default settings that we picked for the bucket and now we can see that default hyphen Merlin.jpg object has been uploaded so if we open up default hyphen Merlin we can see it's using sse-kms as the service side encryption type and it's using the KMS key that we set in the default encryption settings on the bucket.

      Okay well that's everything I wanted to cover in this demo lesson so let's just tidy up to make sure that we don't experience any charges go back to the Amazon S3 console select the bucket that you've created and then click on empty you need to confirm to empty that bucket once that process is completed and the bucket's emptied and then follow that same process but this time click on delete to delete the bucket from your AWS account click on key management service and we'll just mark the key that were created for deletion so select customer managed keys select cat pics click on key actions schedule key deletion set this to 7 which is the minimum check the box and click on schedule deletion with that being said though that's everything I wanted to cover in this demonstration I hope it's been fun and useful go ahead mark this video is complete and when you're ready I'll see you in the next video.

    1. Automatic thinking causes us to simplify problems and see them through narrow frames. We fi ll in miss- ing information based on our assumptions about the world and evaluate situations based on associations that automatically come to mind and belief systems that we take for granted. In so doing, we may form a mistaken picture of a situation, just as looking through a small window overlooking an urban park could mis- lead someone into thinking he or she was in a more bucolic place. page 12

      I think that the results from the research conducted on culting stigmatized identity affecting students' performances made me realize how much of a mental toll stereotypes can play on people. It's disheartening and ironic simultaneously to see how the high and low sides of the caste-system groups collectively performed worse when they were told their respective roles. It influences the way that I think as a student going to an international school because it is interesting to see how these ideas can parallel to students around me. Regardless, this passage relates to today's inquiry question because it can be used and argued to reflect how poor people shape individual economic actions due to neglect and stereotypes affecting their life subconsciously.

    1. Welcome back and in this lesson I want to talk about S3 encryption.

      Now we're going to be focusing on server-side encryption known as SSE, which I will be coaching on client-side encryption and how that's different.

      Now we've got a lot to get through so let's jump in and get started.

      Now before we start there's one common misconception which I want to fix right away, and that's that buckets aren't encrypted, objects are.

      You don't define encryption at the bucket level.

      There's something called bucket default encryption, but that's different and I'll cover that elsewhere in the course.

      For now, understand that you define encryption at the object level, and each object in a bucket might be using different encryption settings.

      Now before we talk about the ways that S3 natively handles encryption for objects, I think it's useful to just review the two main architectures of encryption which can be used with the product.

      There's client-side encryption and server-side encryption, and both of these refer to what method is used for encryption at rest, and this controls how objects are encrypted as they're written to disk.

      It's a method of ensuring that even if somebody were to get the physical disks from AWS which your data is on, they would need something else, a type of key to access that data.

      So visually this is how a transaction between a group of users or an application and S3 looks like.

      The users of the application on the left are loading data to an S3 endpoint for a specific bucket which gets stored on S3's base storage hardware.

      Now it's a simplistic overview, but for this lesson it's enough.

      I want to illustrate the difference between client-side encryption and server-side encryption.

      So on the top we have client-side encryption, and on the bottom we have server-side encryption.

      Now this is a really, really important point which often confuses students.

      What I'm talking about in this lesson is encryption at rest, so how data is stored on disk in an encrypted way.

      Both of these methods also use encryption in transit between the user-side and S3.

      So this is an encrypted tunnel which means that you can't see the raw data inside the tunnel.

      It's encrypted.

      So ignoring any S3 encryption, ignoring how data is encrypted as it's written to disk, data transferred to S3 and from S3 is generally encrypted in transit.

      Now there are exceptions, but use this as your default and I'll cover those exceptions elsewhere in the course.

      So in this lesson when we're talking about S3 encryption, we're focusing on encryption at rest and not encryption in transit, which happens anyway.

      Now the difference between client-side encryption and server-side encryption is pretty simple to understand when you see it visually.

      With client-side encryption, the objects being uploaded are encrypted by the client before they ever leave, and this means that the data is ciphertexted the entire time.

      From AWS's perspective, the data is received in a scrambled form and then stored in a scrambled form.

      AWS would have no opportunity to see the data in its plain text form.

      With server-side encryption known as SSE, it's slightly different.

      Here, even though the data is encrypted in transit using HTTPS, the objects themselves aren't initially encrypted, meaning that inside the tunnel, the data is in its original form.

      Let's assume it's animal images.

      So you could remove the HTTP encrypted tunnel somehow and the animal pictures would be in plain text.

      Now once the data hits S3, then it's encrypted by the S3 servers, which is why it's referred to as server-side encryption.

      So to high level, the differences are with client-side encryption, everything is yours to control.

      You take on all of the risks and you control everything, which is both good and bad.

      You take the original data, you are the only one who ever sees the plain text version of that data, you generate a key, you hold that key and you manage that key.

      You are responsible for recording which key is used for which object, and you perform the encryption process before it's uploaded to S3, and this consumes CPU capacity on whatever device is performing the encryption.

      You just use S3 for storage, nothing else.

      It isn't involved in the encryption process in any way, so you own and control the keys, the process and any tooling.

      So if your organization needs all of these, if you have real reasons that AWS cannot be involved in the process, then you need to use client-side encryption.

      Now with server-side encryption known as SSE, you allow S3 to handle some or all of that process, and this means there are parts that you need to trust S3 with.

      How much of that process you trust S3 with and how you want the process to occur and determine which type of server-side encryption you use as there are multiple types.

      Now AWS has recently made server-side encryption mandatory, and so you can no longer store objects in an unencrypted form on S3.

      You have to use encryption at rest.

      So let's break apart server-side encryption and review the differences between each of the various types.

      There are three types of server-side encryption available for S3 objects, and each is a trade-off of the usual things, trust, overhead, cost, resource consumption and more.

      So let's quickly step through them and look at how they work.

      The first is SSE-C, and this is server-side encryption with customer-provided keys.

      Now don't confuse this with client-side encryption because it's very different.

      The second is SSE-S3, which is server-side encryption with Amazon S3 managed keys, and this is the default.

      The last one is an enhancement on SSE-S3, which is SSE-KMS, and this is server-side encryption with KMS keys stored inside the AWS Key Management Service, known as KMS.

      Now the difference between all of these methods is what parts of the process you trust S3 with and how the encryption process and key management is handled.

      At a high level, there are two components to server-side encryption.

      First, the encryption and decryption process.

      This is the process where you take plain text, a key and an algorithm, and generate cyber text.

      It's also the reverse, so taking that cyber text and a key and using an algorithm to output plain text.

      Now this is symmetrical encryption, so the same key is used for both encryption and decryption.

      The second component is the generation and management of the cryptographic keys, which are used as part of the encryption and decryption processes.

      These three methods of server-side encryption, they handle these two components differently.

      Now let's look at how.

      Now before we do, again, I just want to stress that SSE is now mandatory on objects within S3 buckets.

      This process will occur, you cannot choose not to use it.

      The only thing that you can influence is how the process happens and what version of SSE is utilized.

      Now first, with SSE-C, the customer is responsible for the keys, and S3 manages the encryption and decryption processes.

      So the major change between client-side encryption and this is that S3 are handling the cryptographic operations.

      Now this might sound like a small thing, but if you're dealing with millions of objects and a high number of transactions, then the CPU capability required to do encryption can really add up.

      So you're essentially offloading the CPU requirements of this process to AWS, but you still need to generate and manage the key or keys.

      So when you put an object into S3 using this method, you provide the plain text object and an encryption key.

      Remember this object is encrypted in transit by HTTPS on its way to S3, so even though it's plain text right now, it's not visible to an external observer.

      When it arrives at S3, the object is encrypted and a hash of the key is tagged to the object and the key is destroyed.

      Now this hash is one way, it can't be used to generate a new key, but if a key is provided during decryption, the hash can identify if that specific key was used or not.

      So the object and this one-way hash are stored on disk, assistantly.

      Remember S3 doesn't have the key at this stage.

      To decrypt, you need to provide S3 with the request and the key used to encrypt the object.

      If it's correct, S3 decrypts the object, discards the key and returns the plain text.

      And again, returning the object is done over an encrypted HTTPS tunnel, so from the perspective of an observer, it's not visible.

      Now this method is interesting.

      You still have to manage your keys, which does come with a cost and some effort, but you also retain control of that process, which is good in some regulation-heavy environments.

      You also save on CPU requirements versus client-side encryption, because S3 performs encryption and decryption, meaning smaller devices don't need to consume resources for this process.

      But you need to trust that S3 will discard the keys after use, and there are some independent audits which prove what AWS does and doesn't do during this process.

      So you choose SSE-C when you absolutely need to manage your own keys, but are happy to allow S3 to perform the encryption and decryption processes.

      You would choose client-side encryption when you need to manage the keys and also the encryption and decryption processes, and you might do this if you never want AWS to have the ability to see your plain text data.

      So let's move on to the next type of server-side encryption, and the type I want to describe now is SSE-S3.

      And with this method, AWS handles both the encryption processes as well as the key generation and management.

      When putting an object into S3, you just provide the plain text data.

      When an object is uploaded to S3 using SSE-S3, it's encrypted by a key which is unique for every object, so S3 generates a key just for that object, and then it uses that key to encrypt that object.

      For extra safety, S3 has a key which it manages as part of the service.

      You don't get to influence this, you can't change any options on this key, nor do you get to pick it.

      It's handled end-to-end by S3.

      From your perspective, it isn't visible anywhere in the user interface, and it's rotated internally by S3 out of your visibility and control.

      This key is used to encrypt the per-object key, and then the original key is discarded.

      What we're left with is a ciphertext object and a ciphertext key, and both of these are persistently stored on disk.

      With this method, AWS take over the encryption process just as with SSE-C, but they also manage the keys on your behalf, which means even less admin overhead.

      The flip side with this method is that you have very little control over the keys used.

      The S3 key is outside of your control, and the keys used to encrypt and encrypt objects are also outside of your control.

      For most situations, SSE-S3 is a good default type of encryption which makes sense.

      It uses a strong algorithm, AES256, the data is encrypted at rest and the customer doesn't have any admin overhead to worry about, but it does present three major problems.

      Firstly, if you're in an environment which is strongly regulated, where you need to control the keys used and control access to the keys, then this isn't suitable.

      If you need to control rotation of keys, this isn't suitable.

      And then lastly, if you need role separation, this isn't suitable.

      What I mean by role separation is that a full S3 administrator, somebody who has full S3 permissions to configure the bucket and manage the objects, then he or she can also decrypt and view data.

      You can't stop an S3 full administrator from viewing data when using this type of server-side encryption.

      And in certain industry areas such as financial and medical, you might not be allowed to have this small and open access for service administrators.

      You might have certain groups within the business who can access the data but can't manage permissions, and you might have requirements for another SIS admin group who need to manage the infrastructure but can't be allowed to access data within objects.

      And with SSE-S3, this cannot be accomplished in a rigorous best practice way.

      And this is where the final type of server-side encryption comes in handy.

      The third type of server-side encryption is server-side encryption with AWS Key Management Service Keys, known as SSE-KMS.

      How this differs is that we're now involving an additional service, the Key Management Service, or KMS.

      Instead of S3 managing keys, this is now done via KMS.

      Specifically, S3 and KMS work together.

      You create a KMS key, or you can use the service default one, but the real power and flexibility comes from creating a customer-managed KMS key.

      It means this is created by you within KMS, it's managed by you, and it has isolated permissions, and I'll explain why this matters in a second.

      In addition, the key is fully configurable.

      Now this seems on the surface like a small change, but it's actually really significant in terms of the capabilities which it provides.

      When S3 wants to encrypt an object using SSE-KMS, it has to liaise with KMS and request a new data encryption key to be generated using the chosen KMS key.

      KMS delivers two versions of the same data encryption key, a plain text version and an encrypted or cipher text version.

      S3 then takes the plain text object and the plain text data encryption key and creates an encrypted or cipher text object, and then it immediately discards the plain text key, leaving only the cipher text version of that key and both of these are stored on S3 storage.

      So you're using the same overarching architecture, the per object encryption key, and the key which encrypts the per object key, but with this type of server-side encryption, so using SSE-KMS, KMS is generating the keys.

      Now KMS keys can only encrypt objects up to 4KB in size, so the KMS key is used to generate data encryption keys which don't have those limitations.

      It's important to understand that KMS doesn't store the data encryption keys, it only generates them and gives them to S3.

      But you do have control over the KMS key, the same control as you would with any other customer-managed KMS key.

      So in regulated industries, this alone is enough reason to consider SSE-KMS because it gives fine-grained control over the KMS key being used as well as its rotation.

      You also have logging and auditing on the KMS key itself, and with CloudTrail you'll be able to see any calls made against that key.

      But probably the best benefit provided by SSE-KMS is the role separation.

      To decrypt an object encrypted using SSE-KMS, you need access to the KMS key which was originally used.

      That KMS key is used to decrypt the encrypted copy of the data encryption key for that object which is stored along with that object.

      If you don't have access to KMS, you can't decrypt the data encryption key, so you can't decrypt the object, and so it follows that you can't access the object.

      Now what this means is that if we had an S3 administrator, and let's call him Phil, because we're using SSE-KMS, it means Phil as an S3 administrator does have full control over this bucket.

      But because Phil has been given no permissions on the specific KMS key, he can't read any objects.

      So he can administer the object as part of administering S3, but he can't see the data within those objects because he can't decrypt the data encryption key using the KMS key because he has no permissions on that KMS key.

      Now this is an example of role separation, something which is allowed using SSE-KMS versus not allowed using SSE-S3.

      With SSE-S3, Phil as an S3 administrator could administer and access the data inside objects.

      However, using SSE-KMS, we have the option to allow Phil to view data in objects or not, something which is controllable by granting permissions or not on specific KMS keys.

      So time for a quick summary before we finish this lesson, and it's really important that you understand these differences for any of the AWS exams.

      With client-side encryption, you handle the key management and the encryption and decryption processes.

      Use this if you need to control both of those and don't trust AWS and their regular audits.

      This method uses more resources to manage keys as well as resources for actually performing the encryption and decryption processes at scale.

      But it means AWS never see your objects in plain text form because you handle everything end to end.

      This generally means you either encrypt all objects in advance or use one of the client-side encryption SDKs within your application.

      Now please don't confuse client-side encryption with server-side encryption, specifically SSE-C.

      Client-side encryption isn't really anything to do with S3, it's not a form of S3 encryption, it's different.

      You can use client-side encryption and server-side encryption together, there's nothing preventing that.

      So now let's step through server-side encryption, and remember this is now on by default, it's mandatory.

      The only choice you have is which method of SSE to use.

      With SSE-C you manage the encryption keys, you can use the same key for everything, but that isn't recommended.

      Or you can use individual keys for every single object, the choice is yours.

      S3 accepts your choice of key and an object and it handles the encryption and decryption processes on your behalf.

      This means you need to trust S3 with the initial plain text object and trust it to discard and not store the encryption key.

      But in exchange S3 takes over the computationally heavy encryption and decryption processes.

      And also keep in mind that the data is transferred in a form where it's encrypted in transit using HTTBS.

      So nobody outside AWS will ever have exposure for plain text data in any way.

      SSE-S3 uses AES-256, I mention this because it's often the way exam questions test your knowledge.

      If you see AES-256, think SSE-S3.

      With SSE-S3, S3 handles the encryption keys and the encryption process.

      It's the default and it works well for most cases, but you have no real control over keys, permissions or rotation.

      And it also can't handle role separation, meaning S3 for admins can access the data within objects that they manage.

      Finally we have SSE-KMS which uses KMS and KMS keys which the service provides.

      You can control key rotation and permissions, it's similar in operation to SSE-S3, but it does allow role separation.

      So use this if your business has fairly rigid groups of people and compartmentalised sets of security.

      You can have S3 admins with no access to the data within objects.

      Now for all AWS exams make sure you understand the difference between client side and server side encryption.

      And then for server side encryption try and pitch scenarios where you would use each of the three types of server side encryption.

      Now that's everything I wanted to cover in this lesson about object encryption, specifically server side encryption.

      Go ahead and complete this lesson, but when you're ready I look forward to you joining me in the next.

    1. Welcome back and in this demo lesson I just wanted to give you a bit of practical exposure to KMS.

      And in this demo we're going to continue using the scenario that we've got the cat ruler trying to send encrypted battle plans to the robot general.

      Now to get started just make sure that you're logged in into the IAM admin user of the general AWS account, so the management account of the organisation.

      As always you'll need to have the Northern Virginia region selected and once you do go ahead and move across to the KMS console.

      So click in the search box at the top, type KMS and then open that in a new tab.

      And then click on create key because we'll be creating a KMS key.

      Now KMS allows the creation of symmetric or asymmetric keys and just to keep this demonstration simple we're going to demonstrate using a symmetric key.

      So make sure this option selected and then just to demonstrate some of these options just expand this and this is where you can set the key material origin.

      If you recall from the previous lesson I talked about how the physical backing key material could be generated by KMS or imported and this is where you select between those two options.

      Now I'm not going to talk about the custom key store at this point in the course.

      I'll talk about this in more detail later on in the course when I talk about cloud HSM which is a different product entirely.

      For now just select KMS and this will make KMS generate this physical backing key material that we'll be using to do our cryptographic operations.

      Now historically KMS was a single region service which means that keys created in the product could never leave the region that they were created in.

      More recently KMS has extended this functionality allowing you to create multi region keys.

      Now for the purpose of this demo lesson we're only going to be using a single region key which is the default.

      So make sure that single region key is selected and then we can continue to go ahead and click on next.

      Now in the previous lesson I mentioned how a key has a unique ID but also that we could create an alias which references the key.

      And that's what we can do here so I'm going to go ahead and create an alias and I'm going to call the alias cat robot all onward.

      So type in cat robot and click next.

      Now I discussed earlier how a KMS key has a key policy and a key policy is a resource policy that applies to the key.

      Now it's here where we can specify the key administrators for a particular key.

      There is a difference between identities that can manage a key and identities that can use a key for cryptographic operations like encrypt or decrypt.

      It's this point where we define who can manage this key.

      So go ahead and just check the box next to I am admin.

      So that will make sure that our I am admin user can administer this key.

      Once you do that scroll down there's a box here that allows administrators also to delete this key and that's a default so we'll leave that as is and click next.

      Now the previous set is where we defined who had admin permissions on the key.

      This stage lets us define who can use the key for cryptographic operations so encrypt and decrypt.

      To keep things simple we're going to define a key which adds the relevant entries to the key policy.

      So just check the box next to I am admin which is the user that we logged in as and then just scroll down.

      So if you wanted to add other AWS accounts here so that they had permission to use this key but for this demonstration we don't need to do that.

      So just click on next and this is the key policy that this wizard has created.

      So if I just scroll down it assigns the account level trust at the top so the account itself, the account user is allowed to perform any KMS column actions on this key.

      So that is the part of the policy of this statement here which means that this key will trust this account.

      It's this statement that defines the key administrators so that I am admin inside this account can create and describe and enable and list and all the other admin style permissions.

      Scroll down further still it's this statement that allows I am admin to perform encryption and decrypt and re-encrypt and generate data key and describe key actions against this key.

      So all the permissions that we define are inside the key policy.

      So this point go ahead click on finish and that will create the key as well as the alias that we use to reference this key.

      If we go into this key I'll just show you some of the options that are available.

      We'll be able to obviously edit the key policy and we can define key rotation.

      So by default key rotation for a customer managed key is switched off and we can enable it to rotate the key once every year.

      For an AWS managed key that is by default turned on and you can't switch it off and it also performs a rotation approximately once every year.

      So just click on AWS managed keys as we go through the course and start turning on encryption for various different services.

      You'll notice how each service the first time it uses encryption with KMS it creates an AWS managed key in this list.

      Now that's everything we need to do on the AWS side.

      Now we can start using this key to perform some cryptographic operations.

      So let's do that.

      Now at this point rather than using the local command line interface on your local machine we're going to be using Cloud Shell.

      This allows us to use the same set of instructions regardless of your local operating system.

      So to launch Cloud Shell click on this icon and this will take a few minutes but it will put you at the shell that's using your currently logged in user in order to gain permission.

      So any commands you run in the shell will be run as your currently logged in user.

      So the first thing that we'll be doing is to create a plain text battle plan.

      So this is the message that the cab ruler is going to be sending to the robot general.

      To generate that file we'll use echo and then space and then a speechmark and then a small message and the message is going to be find all the doggos and then a comma distract them with the yums.

      So find all the doggos, distract them with the yums and then a speechmark to close that off and then we'll redirect that to a file called battleplans.txt.

      And then presenter.

      Now the commands to interact with KMS from the command line are fairly long so what I'll do is paste it in and then I'll step through it line by line and explain exactly what it's doing.

      So first we need to encrypt the plain text battle plans and we want the result to be a cipher text document something that we can pass to the robot general which can't be intercepted en route and can be decrypted at the other end.

      So this is the command that we need to run and I just want to step through this line by line.

      The top part should be fairly obvious so we're running the AWS command line tools, the KMS module and using the encrypt action.

      So this specifies that we want to encrypt piece of data.

      This line specifies the alias that we want to use to encrypt this piece of data.

      You can either specify the key ID which uniquely identifies a key or you can specify an alias using the alias forward slash and then the name of the alias.

      In this case I've lept it to do that so this is using the alias that we created in the first part of this demo lesson.

      This line is where we're providing the plain text to KMS and instead of typing the plain text on the command line we're telling it to consult this file so battleplans.txt.

      Now the next line is telling the command line tools to output the result as text and it's going to be a text output with a number of different fields.

      The next line, double hyphen query, is telling the command line tools to select one particular field and that's the field cipher text blob and it's this field that contains the cipher text output from the KMS command.

      Now the output of any of these commands that interact with KMS is going to be a base64 encoded file so it's not going to be binary data, it's going to be base64 encoded.

      What we want to do is have our output being a binary encrypted file and so we need to take the result of this encryption command and pass it to a utility called base64 and that utility using this command line option will decode that base64 and place the result into a not_battleplans.enc file and this is going to be our result in cipher text.

      Now I know that command is relatively complex, KMS is not the easiest part of AWS to use from the command line but I did want to step you through line by line so you didn't know what each line achieved.

      Ok so let's go ahead and run this command, to do that we need to click on paste and then once that's pasted into a cloud shell press enter to run the command and the output not_battleplans will be our encrypted cipher text.

      So if I run a cat not_battleplans we get binary encrypted data so obviously anyone looking from the outside will just see scrambled data and won't understand what the message is.

      So now I'm going to clear the screen to make it a little bit easier to see and this is the encrypted cipher text file that we could transfer across to the robot general.

      So now we need to assume in this scenario that we're now the robot general and we're looking to decrypt this file.

      Ok so now I'm going to paste the next command for this lesson which is the decrypt command and I'll be stepping through line by line just explaining exactly what each line accomplishes.

      So this is the command that you use to decrypt the cipher text and give us the original plain text battle plans.

      So first this top line should be logical we're running the AWS command line tools with the KMS module and the decrypt command.

      We're passing in some cipher text so we use the command line option double hyphen cipher text blob and instead of pasting this on the command line we're giving it this file so not_battleplans.enc.

      We're again asking for the output to be in text.

      This will output some text with a number of different fields we're using the query field query for the plain text field and again the output will be base 64 encoded and so we're using the base 64 utility with the double hyphen decode to decode that back into its original form and store that into a file called decryptorplans.txt.

      So let's go ahead and run this so click paste and then press enter to run this command.

      This will decrypt cipher text and it will output decryptorplans.txt.

      And if we catch that document we'll see the original message.

      Find all the doggos, distract them with the yums and that's just been a really simple demonstration of using the KMS encrypt command and the KMS decrypt command.

      A couple of things I wanted to draw your attention to throughout the process.

      With the encrypt command we needed to pass in the key to use as well as the plain text and we got out the cipher text.

      With the decrypt command we don't need to specify the key, we only give the cipher text and assuming we have permissions on the KMS key so that we can use it to perform decrypt operations then we'll get the decryptor plain text and that's what's happened here.

      Now just to clear up from this lesson if you go back to the AWS console make sure you're in US East 1 so Northern Virginia and go back to the key management service console and we're just going to delete the KMS key that we created earlier in this lesson.

      So click on customer managed keys, select the KMS key that we created earlier, my case, cap robot then click on key actions and schedule key deletion.

      You need to enter a waiting period between 7 and 30 days since you want this cleared up as fast as possible going into 7, tick the box to confirm and then schedule delete.

      And that'll put the key into a pending deletion state and after 7 days it'll be entirely removed.

      And at that point we've cleared up all of the assets that we've used in this demo lesson so go ahead and complete the video and when you're ready join me in the next.

    1. Welcome to this video where I'm going to be talking about the key management service known as KMS.

      Now this product is used by many other services within AWS when they use encryption, so don't be surprised if you're watching this video in what seems like a pretty random place within the course.

      With that being said let's jump in and get started.

      Now KMS isn't all that complex as a product.

      Once you understand it it's pretty simple, but because of how much it's used by other AWS products and services it's essential that you do understand it for all the AWS exams.

      Now KMS is a regional and public service.

      Every region is isolated when using KMS.

      Think of it as a separate product.

      Now KMS is capable of some multi-region features but I'll be covering those in a separate dedicated video.

      It's a public service which means it occupies the AWS public zone and can be connected to from anywhere with access to this zone.

      Like any other AWS service though you will need permissions to access it.

      Now KMS as the name suggests manages keys.

      Specifically it lets you create, store and manage cryptographic keys.

      These are keys which can be used to convert plain text to ciphertext and vice versa.

      Now KMS is capable of handling both symmetric and asymmetric keys.

      And at this point you should understand what that means.

      Where symmetric keys are used, where public asymmetric are used, as well as private asymmetric.

      Just know that KMS is capable of operating with all of these different key architectures.

      Now KMS is also capable of performing cryptographic operations which includes, but is not limited to, encryption and encryption operations.

      And I'll be talking more about this later in this video.

      Now one of the foundational things to understand about KMS is that cryptographic keys never leave the product.

      KMS can create keys, keys can be imported, it manages keys, it can use these keys to perform operations but the keys themselves are locked inside the product.

      Its primary function is to ensure the keys never leave and held securely within the service.

      Now KMS also provides a FIPS 140-2 compliant service.

      This is a US security standard but try to memorize this.

      It's FIPS 140-2 level 2 to be specific.

      Again the level 2 part matters.

      It's often a key point of distinction between using KMS versus using something like cloud HSM which I'll be covering in detail elsewhere.

      Now some of KMS's features have achieved level 3 compliance but overall it's level 2.

      Again please do your best to remember this.

      It will come in handy for most of the AWS exams.

      Now before we continue, since this is an introduction video, unless I state otherwise, assume that I'm talking about symmetric keys.

      When I mention keys within this video, I'm going to be covering the advanced functionality of KMS in other videos including asymmetric keys but for this one I'm mainly focusing on its architecture and high level functions.

      So just assume I'm talking about symmetric keys from now on unless I indicate otherwise.

      Now the main type of key that KMS manages are known logically enough as KMS keys.

      You might see these referred to as CMKs or Customer Master Keys but that naming scheme has been superseded so they're now called KMS keys.

      These KMS keys are used by KMS within cryptographic operations.

      You can use them, applications can use them and other AWS services can use them.

      Now they're logical, think of them as a container for the actual physical key material and this is the data that really makes up the key.

      So a KMS key contains a key ID, a creation date, a key policy which is a resource policy, a description and a state of the key.

      Every KMS key is backed by physical key material, it's this data which is held by KMS and it's this material which is actually used to encrypt and decrypt things that you give to KMS.

      The physical key material can be generated by KMS or imported into KMS and this material contained inside a KMS key can be used to directly encrypt or decrypt data up to 4KB in size.

      Now this might sound like a pretty serious limitation.

      KMS keys are generally only used to work on small bits of data or to generate other keys and I'll be covering this at a high level later in this video.

      Let's look visually at how KMS works so far.

      So this is KMS and this is Ashley.

      Ashley's first interaction with KMS after picking a region is to create a KMS key.

      A KMS key is created with physical backing material and this key is stored within KMS in an encrypted form.

      Nothing in KMS is ever stored in plain text form persistently.

      It might exist in memory in plain text form but on disk it's encrypted.

      Now Ashley's next interaction with KMS might be to request that some data is encrypted.

      To do this she makes an encrypt call to KMS specifying the key to use and providing some data to encrypt.

      KMS accepts the data and assuming Ashley has permissions to use the key, it decrypts the KMS key then uses this key to encrypt the plain text data that Ashley supplied and then returns that data to Ashley.

      Notice how KMS is performing the cryptographic operations.

      Ashley is just providing data to KMS together with instructions and it's handling the operations internally.

      Logically at some point in the future Ashley will want to decrypt this same data so she calls a decrypt operation and she includes the data she wants to decrypt along with this operation.

      KMS doesn't need to be told which KMS key to use for the decrypt operation.

      That information is encoded into the cyber text of the data which Ashley wants to decrypt.

      The permissions to decrypt are separate from the permissions to encrypt and are also separate from permissions which allow the generation of keys but assuming Ashley has the required permissions for a decrypt operation using this specific KMS key, KMS decrypts the key and uses this to decrypt the cyber text provided by Ashley and returns this data back to Ashley in plain text form.

      Now again I want to stress at no point during this entire operation do the KMS keys leave the KMS product.

      At no point are the keys stored on the disk in plain text form and at each step Ashley needs permissions to perform the operations and each operation is different.

      KMS is very granular with permissions.

      You need individual permissions for various operations including encrypt and decrypt and you need permissions on given KMS keys in order to use those keys.

      Ashley could have permissions to generate keys and to use keys to encrypt and decrypt or she could have just one of those permissions.

      She might have permissions to encrypt data but not decrypt it or she might have permissions to manage KMS creating keys and setting permissions but not permissions to use keys to encrypt or decrypt data and this process is called "Rowl Separation".

      Now I mentioned at the start of this lesson that the KMS key can only operate cryptographically on data, which is a maximum of 4kb in size.

      Now that's true, so let's look at how KMS gets around this.

      Data encryption keys, also known as DEX or D-E-Ks, are another type of key which KMS can generate.

      They're generated using a KMS key, using the generate data key operation.

      This generates a data encryption key which can be used to encrypt and decrypt data which is more than 4kb in size.

      Data encryption keys are linked to the KMS key which created them, so KMS can tell that a specific data encryption key was created using a specific KMS key.

      But, and this is pretty much the most important thing about KMS and data encryption keys, KMS doesn't store the data encryption key in any way.

      It provides it to you or the service using KMS and then it discards it.

      The reason it discards it is that KMS doesn't actually do the encryption or decryption of data using data encryption keys.

      You do or the service using KMS performs those operations.

      So let's look at how this works.

      The data encryption key is generated, KMS provides you with two versions of that data encryption key.

      First, a plain text version of that key, something which can be used immediately to perform cryptographic operations.

      And second, a ciphertext or encrypted version of that same data encryption key.

      The data encryption key is encrypted using the KMS key that generated it.

      And in future, this encrypted data encryption key can be given back to KMS for it to be decrypted.

      Now the architecture is that you would generate a data encryption key immediately before you wanted to encrypt something.

      You would encrypt the data using the plain text version of the data encryption key and then once finished with that process, discard the plain text version of that data encryption key.

      That would leave you with the encrypted data and you would then store the encrypted data encryption key along with that encrypted data.

      Now a few key things about this architecture.

      KMS doesn't actually do the encryption or decryption on data larger than 4KB using data encryption keys.

      You do or the service using KMS does.

      KMS doesn't track the usage of data encryption keys.

      That's also you or the service using KMS.

      You could use the same data encryption key to encrypt 100 or a million files or you could request a new data encryption key for each of those million files.

      How you decide to do this is based on your exact requirements and of course AWS services will make this choice based on their requirements.

      By storing the encrypted data encryption key on disk with the encrypted data, you always have the correct data encryption key to use.

      But both the deck and the data are encrypted so administration is easy and security is maintained.

      When you're encrypting that data is simple.

      You pass the encrypted data encryption key back to KMS and ask for it to decrypt it using the same KMS key used to generate it.

      Then you use the decrypted data encryption key that KMS gives you back and decrypt the data with it and then you discard the decrypted data encryption key.

      Services such as S3 when using KMS generate a data encryption key for every single object.

      They encrypt the object and then discard the plain text version.

      As we move through the course I'll be talking in detail about how those services integrate with KMS for encryption services.

      Before we finish up with this lesson there are a few key concepts which I want to discuss.

      The one thing which is really important to grasp with KMS is that by default KMS keys are stored within the KMS service in that specific region.

      They never leave the region and they never leave the KMS service.

      You cannot extract a KMS key.

      Any interactions with a KMS key are done using the APIs available from KMS.

      Now this is the default but KMS does support multi-region keys where keys are replicated to other AWS regions.

      But I'll be covering that in a dedicated video if required for the course that you're studying.

      In KMS as a product keys are either AWS owned or customer owned.

      We're going to be dealing mainly with customer owned keys.

      AWS owned keys are a collection of KMS keys that an AWS service owns and manages for use in multiple AWS accounts.

      They operate in the background and you largely don't need to worry about them.

      If applicable for the course that you're studying I'll have a separate video on this.

      If not don't worry it's unimportant.

      Now in dealing with customer owned keys there are two types.

      AWS managed and customer managed and I'll be covering the specifics of these in a dedicated video.

      AWS managed keys are created automatically by AWS when you use a service such as S3 which integrates with KMS.

      Customer managed keys are created explicitly by the customer to use directly in an application or within an AWS service.

      Customer managed keys are more configurable.

      For example you can edit the key policy which means you could allow cross account access so that other AWS accounts can use your keys.

      AWS managed keys can't really be customized in this way.

      Both types of keys support rotation.

      Rotation is where physical backing materials are the data used to actually do cryptographic operations is changed.

      With AWS managed keys this can't be disabled.

      It's set to rotate approximately once per year.

      With customer managed keys rotation is optional.

      It's enabled by default and happens approximately once every year.

      A KMS key contains the backing key, the physical key material and all previous backing keys caused by rotation.

      It means that as a key is rotated data encrypted with all versions can still be decrypted.

      Now you can create aliases which are shortcuts to keys.

      So you might have an alias called my app one which points at a specific KMS key.

      That way KMS keys can be changed if needed.

      But be aware the aliases are also per region.

      You can create my app one in all regions but in each region it will point at a different key.

      Neither aliases or keys are global by default.

      Okay to finish up this KMS 101 lesson I want to talk at high level about permissions on KMS keys.

      Permissions on keys are controlled in a few ways.

      KMS is slightly different than other AWS services that you come across in terms of how keys are handled.

      Many services will always trust the account that they're contained in.

      Meaning if you grant access via an identity policy that access will be allowed unless there's an explicit deny.

      KMS is different.

      This account trust is explicitly added on a key policy or not.

      The starting point for KMS security is the key policy.

      This key policy is a type of resource policy like a bucket policy only on a key.

      Every KMS key has one and for custom managed keys you can change it.

      To reiterate this the reason the key policy is so important is that unlike other AWS services KMS has to explicitly be told that keys trust the AWS account that they're contained within.

      And this is what a key policy might look like.

      It means that the key will allow the account 11112222333 to manage it.

      This trust isn't automatic so be careful when updating it.

      You always need this type of key policy in place if you want to be able to grant access to a key using identity policies.

      The key doesn't trust the AWS account and this means that you would need to explicitly add any permissions on the key policy itself.

      Generally KMS is managed using this combination of key policies trusting the account and then using identity policies to let IAM users interact with the key.

      But in high security environments you might want to remove this account trust and insist on any key permissions being added inside the key policy.

      And a typical IAM permissions policy for KMS might look something like this which gives the holder of the policy the rights to use this key to encrypt or decrypt data.

      Inside KMS permissions are very granular and can be split based on function.

      You can be granted rights to create keys and manage keys but not to have permissions to perform cryptographic operations like encrypt or decrypt.

      This way your product administrators are given rights to access data encrypted by KMS which is a common requirement of many higher security environments.

      Now there's another way to interact with KMS using grants but I'll be covering this elsewhere in another video if needed.

      So that's everything I wanted to cover in this KMS introduction video.

      This video is going to form the foundation for others in this series depending on the topic that you're studying there might be no more videos or many more videos.

      Don't be worried it'll be the case.

      Now at this point that's everything I wanted to talk about though about KMS at this introductory level.

      Go ahead and complete the video and when you're ready I look forward to you joining me in the next.

    1. Welcome back.

      In this demo lesson, I just want to give you a really quick overview of the S3 performance improvement feature that we talked about in the previous lesson.

      Now, I want to make this really brief because it's not something that I can really demonstrate effectively because I have a relatively good internet connection.

      So I'm not going to be able to achieve the differences in performance that will effectively demonstrate this feature.

      But we can see how to enable it and then use an AWS tool to see exactly what benefits we can expect.

      So to enable it, we can move to the S3 console and just create a bucket.

      Now, you won't be able to use bucket names with periods in with this accelerator transfer tool.

      So this would not be a valid bucket name to use.

      It cannot have periods in the name.

      So I'm going to go ahead and create a test bucket.

      I'm going to call it test AC1279, which is something I don't believe I've used before.

      Now, you should go ahead and create a bucket yourself just to see where to enable this, but you can just follow along with what I'm doing.

      This won't be required at any point further down the line in the course.

      So if you just want to watch me in this demo, that's completely okay.

      So I'll create this bucket.

      I'll accept the rest of the defaults.

      Then I'll select the bucket and go to properties and it's right down at the bottom where you enable transfer acceleration.

      So this is an on/off feature.

      So if I select it, I can enable transfer acceleration or disable it.

      And when I enable it, it provides a new endpoint for this bucket to utilize the transfer acceleration feature.

      So that's important.

      You need to use this specific endpoint in order to get the benefit of these accelerator transfers.

      And this will resolve to an edge location that is highly performant wherever you're physically located in the world.

      So it's important to understand that you need to use this endpoint.

      So enable this on the bucket and click on save and that's all you need to do.

      That enables the feature.

      Now, as I mentioned at the start of this lesson, this is not the easiest feature to demonstrate, especially if you have a good internet connection.

      This is much better as a demonstration if I was on a suboptimal connection, which I'm not right now.

      But what you can do is click on the URL which is attached to this lesson, which opens a comparison tool.

      So what this is going to do now is it's going to compare the direct upload speeds that I can achieve from my local machine to this specific S3 bucket.

      And it's going to do so with and without transfer acceleration.

      So it will be giving you a comparison of exactly what speed you can expect to upload to a specific AWS region, such as US East 1, and then how that compares, uploading to that same region using accelerator transfer.

      So we can see already this is demonstrating a 144% faster upload speed from my location to Northern Virginia by using accelerator transfer.

      Now, I'm going to allow this to continue running while I continue talking, because what you'll see if you run this tool is a different set of results than what I'm seeing.

      You'll see different benefits in each region of using accelerator transfer, depending on your home location.

      So if you're located in San Francisco, for example, you probably won't see a great deal of difference between directly uploading and using accelerator transfer.

      But for more distant regions, you'll see a much more pronounced improvement.

      So if I just move down on this page and make these different regions a little bit easier to see, you'll note, for example, I achieve a much larger benefit in Oregon than I do in San Francisco.

      And my result for Dublin, which is even further away from my current location, is a yet higher benefit for using accelerator transfer.

      So the less optimal the network route is between your location and the region that's being tested, the better benefit you'll achieve by using accelerator transfer.

      Now, there are quite a lot of AWS regions, so I'm not going to let this test finish, but I do recommend if you are interested in S3 performance and this feature specifically, you should test this from your internet connection and allow this process to finish, because it will give you a really good indication of what performance you can expect to each of the AWS regions, and then how S3 accelerator transfer will improve that performance.

      Now, that is everything I wanted to cover in this lesson.

      I know it's been a brief demo lesson, and it isn't really a demo lesson where you're doing anything practically, but I did just want to supplement the previous lesson by giving you a visual example of how this feature can improve performance to S3.

      And I do hope you'll try this tool from your internet connection so you can see the benefit it provides from your location.

      With that being said, though, that is everything that I wanted to cover, so go ahead, complete this video, and when you're ready, I'll see you in the next.

    1. Welcome back, and this time we're going to cover a few performance optimization aspects of S3.

      If you recall from earlier in the course, this is the Animals For Life scenario.

      We have a head office in Brisbane, remote offices which consume services from the Brisbane office, and remote workers using potentially slower or less reliable services to access and upload data to and from the head office.

      So keep the scenario in mind as we step through some of the features that S3 offers to improve performance.

      It's not always about performance.

      It's often about performance and reliability combined.

      And this is especially relevant when we're talking about a distributed organization such as Animals For Life.

      So let's go ahead and review the features that S3 offers, which help us in this regard.

      Now, understanding the performance characteristics of S3 is essential as a solutions architect.

      We know from the Animals For Life scenario that remote workers need to upload large data sets and do so frequently.

      And we know that they're often on unreliable internet connections.

      Now, this is a concern because of the default way that S3 uploads occur.

      By default, when you upload an object to S3, it's uploaded as a single blob of data in a single stream.

      A file becomes an object, and it's uploaded using the put object API call and placed in a bucket.

      And this all happens as a single stream.

      Now, this method has its problems.

      While it is simple, it means that if a stream fails, the whole upload fails, and the only recovery from it is a full restart of the entire upload.

      If the upload fails at 4.5 GB of a 5 GB upload, that's 4.5 GB of data wasted and probably a significant amount of time.

      Remember, the data sets are being uploaded by remote workers over slow and potentially unreliable internet links.

      And this data is critical to the running of the organization.

      Any delay can be costly and potentially risky to animal welfare.

      When using this single put method, the speed and reliability of the upload will always be limited because of this single stream of data.

      If you've ever downloaded anything online, it's often already using multiple streams behind the scenes.

      There are many network-related reasons why even on a fast connection, one stream of data might not be optimal, especially if the transfer is occurring over long distances.

      In this type of situation, single stream transfers can often provide much slower speeds than both ends of that transfer are capable of.

      If I transfer you data with a single stream, it will often run much slower than my connection can do and your connection can do.

      Remember, when transferring data between two points, you're only ever going to experience the speed, which is the lowest of those two points, but often using single stream transfer, you don't even achieve that.

      Data transfer protocols such as BitTorrent have been developed in part to allow speedy distributed transfer of data.

      And these have been designed to address this very concern.

      Using data transfer with only a single stream is just a bad idea.

      Now, there is a limit within AWS if you utilize a single put upload, then you're limited to 5 GB of data as a maximum.

      But I would never trust a single put upload with anywhere near that amount of data.

      It's simply unreliable.

      But there is a solution to this.

      And that solution is multi-part upload.

      Multi-part upload improves the speed and reliability of uploads to S3.

      And it does this by breaking data up into individual parts.

      So we start with the original blob of data that we want to upload to S3, and we break this blob up into individual parts.

      Now, there is a minimum.

      The minimum size for using multi-part upload is 100 MB.

      So the minimum size for this original blob of data is 100 megabytes.

      You can't use multi-part upload if you're uploading data smaller than this.

      Now, my recommendation is that you start using this feature the second that you can.

      The most AWS tooling will automatically use it as soon as it becomes available, which is at this 100 MB lower threshold.

      There are almost no situations where a single put upload is worth it when you get above 100 MB.

      The benefits of multi-part upload are just too extensive and valuable.

      Now, an upload can be split into a maximum of 10,000 parts.

      And each part can range in size between 5 MB and 5 GB.

      The last part is left over, so it can be smaller than 5 MB if needed.

      Now, the reason why multi-part upload is so effective is that each individual part is treated as its own isolated upload.

      So each individual part can fail in isolation and be restarted in isolation, rather than needing to restart the whole thing.

      So this means that the risk of uploading large amounts of data to S3 is significantly reduced.

      But not only that, it means that because we're uploading lots of different individual bits of data, it improves the transfer rate.

      The transfer rate of the whole upload is the sum of all of the individual parts.

      So you get much better transfer rates by splitting this original blob of data into smaller individual parts and then uploading them in parallel.

      It means that if you do have any single stream limitations on your ISP or any network inefficiencies by uploading multiple different streams of data, then you more effectively use the internet bandwidth between you and the S3 endpoint.

      Now, next, I want to talk about a feature of S3 called Accelerated Transfer.

      To understand Accelerated Transfer, it's first required to understand how global transfer works to S3 buckets.

      Let's use an example.

      Let's say that the Animals for Life Organization has a European campaign which is running from the London office.

      For this campaign, there'll be data from staff in the field.

      And let's say that we have three teams dedicated to this campaign, one in Australia, one in South America, and one on the West Coast of the US.

      Now, the S3 bucket, which is being used by the campaign staff, has been created in the London region.

      So this is how this architecture locks.

      We've got three geographically spread teams who are going to be uploading data to an S3 bucket that's located within the UK.

      Now, it might feel like when you upload data to S3, your data would go in a relatively straight line, the most efficient line to its destination.

      Now, this is not how networking works.

      How networking works is that it is possible for the data to take a relatively indirect path.

      And the data can often slow down as it moves from hop to hop on the way to its destination.

      In some cases, the data might not be routed the way you expect.

      I've had data, for instance, routed from Australia to the UK, but taking the alternative path around the world.

      It's often not as efficient as you expect.

      Remember, S3 is a public service, and it's also regional.

      In the case of the Australian team, their data would have to transit across the public internet all the way from Australia to the UK before it enters the AWS public zone to communicate with S3.

      And we have no control over the public internet data path.

      Routers and ISPs are picking this path based on what they think is best and potentially commercially viable.

      And that doesn't always align with what offers the best performance.

      So using the public internet for data transit is never an optimal way to get data from source to destination.

      Luckily, as Solutions Architects, we have a solution to this, which is S3 transfer acceleration.

      Transfer acceleration uses the network of AWS edge locations, which are located in lots of convenient locations globally.

      An S3 bucket needs to be enabled for transfer acceleration.

      The default is that it's switched off, and there are some restrictions for enabling it.

      The bucket name cannot contain periods, and it needs to be DNS compatible in its naming.

      So keep in mind those two restrictions.

      But assuming that's the case, once enabled, data being uploaded by our field workers, instead of going back to the S3 bucket directly, it immediately enters the closest, best performing AWS edge location.

      Now this part does occur over the public internet, but geographically, it's really close, and it transits through fewer normal networks, so it performs really well.

      At this point, the edge locations transit the data being uploaded over the AWS global network, a network which is directly under the control of AWS, and this tends to be a direct link between these edge locations and other areas of the AWS global network, in this case, the S3 bucket.

      Remember, the internet is a global, multi-purpose network, so it has to have lots of connections to other networks, and many stops along the way, where traffic is routed from network to network, and this just slows performance down.

      Think of the internet as the normal public transit network, when you need to transit from bus to train to bus to bus, to get to a far-flung destination.

      The normal transit network, whilst it's not the highest performance, is incredibly flexible, because it allows you to get from almost anywhere to almost anywhere.

      The internet is very much like that.

      It's not designed primarily for speed.

      It's designed for flexibility and resilience.

      The AWS network, though, is purpose-built to link regions to other regions in the AWS network, and so this is much more like an express train, stopping at only the source and destination.

      It's much faster and with lower consistent latency.

      Now, the results of this, in this context, are more reliable and higher performing transfers between our field workers and the S3 bucket.

      The improvements can vary, but the benefits achieved by using transfer acceleration improve the larger the distance between the upload location and the location of the S3 bucket.

      So in this particular case, transferring data from Australia to a bucket located in Europe, you'll probably see some significant gains by using transfer acceleration.

      The worse the initial connection, the better the benefit by using transfer acceleration.

      Okay, so now it's time for a demonstration.

      In the next lesson, I just want to take a few moments to show you an example of how this works.

      I want to show you how to enable the feature on an S3 bucket, and then demonstrate some of the performance benefits that you can expect by using an AWS-provided tool.

      So go ahead, finish this video, and when you're ready, you can join me in the demo lesson.

    1. Did it work prior to replacing the ribbon? If yes, then perhaps remove the ribbon and replace again. See page 19 of the manual here: https://site.xavier.edu/polt/typewriters/RoyalKMM.pdf

      YouTube also has tutorials for how to thread these. (Also search for the No. 10, KH, KHM, HH, Empress, FP, etc. which also used the same general ribbon spools and set up if you can't find a KHM.) I can*t tell 100% from the photo, but the ribbon looks like it's spooling on clockwise on the right (and vice-versa for the left) and you want it the other way.

      Is it not advancing regardless of which direction you have the ribbon going? Usually just one side is not working. You can use this fact to compare the typewriter bilaterally. Watch what's going on with the side that does work and compare it with the side the doesn't. What's wrong on the non-working side?

      Often times the spindle on one or both sides is frozen up with dried up grease, oil, dirt, or dust. A small quirt of mineral spirits or lacquer thinner (or other degreaser) will free it up. (Here we use the mantra, a typewriter isn't really "broken" unless it's clean and broken.) See: https://boffosocko.com/2024/08/09/on-colloquial-advice-for-degreasing-cleaning-and-oiling-manual-typewriters/

      reply to u/UltimateAiden98 at https://old.reddit.com/r/typewriters/comments/1f0nzt8/my_royal_kmm_ribbon_is_t_advancing_what_should_i/

    1. Welcome back and in this lesson I'm going to cover object versioning and MFA delete, two essential features of S3.

      These are two things I can almost guarantee will feature on the exam and almost every major project I can involved in has needed solid knowledge of both.

      So let's jump in and get started.

      Object versioning is something which is controlled at a bucket level.

      It starts off in a disabled state.

      You can optionally enable versioning on a disabled bucket, but once enabled you cannot disable it again.

      Just to be super clear, you can never switch bucket versioning back to disabled once it's been enabled.

      What you can do though is suspend it and if desired a suspended bucket can be re-enabled.

      It's really important for the exam to remember these stage changes.

      So make a point of noting them down and when revising try to repeat until it sticks.

      So a bucket starts off as disabled, it can be re-enabled again, an enabled bucket can be moved to suspended and then moved back to enabled.

      But the important one is that enabled bucket can never be switched back to disabled.

      That is critical to understand for the exam.

      So you can see many trick questions which will test your knowledge on that point.

      Without versioning enabled on a bucket, each object is identified solely by the object key, its name, which is unique inside the bucket.

      If you modify an object, the original version of that object is replaced.

      Versioning lets you store multiple versions of an object within a bucket.

      Any operations which would modify an object, generate a new version of that object and leave the original one in place.

      For example, let's say I have a bucket and inside the bucket is a picture of one of my cats, Winky.

      So the object is called Winky.JPEG.

      It's identified in the bucket by the key, essentially its name, and the key is unique.

      If I modify the Winky.JPEG object or delete it, those changes impact this object.

      Now there's an attribute of an object which I haven't introduced yet and that's the ID of the object.

      When versioning on a bucket is disabled, the ID of the object in that bucket are set to null.

      That's what versioning being off on a bucket means.

      All of the objects have an ID of null.

      Now if you upload or put a new object into a bucket with versioning enabled, then S3 allocates an ID to that object.

      In this case, 111, 111.

      If any modifications are made to this object, so let's say somebody accidentally overrides the Winky.JPEG object with the dog picture, but still calls it Winky.JPEG.

      S3 doesn't remove the original object.

      It allocates a new ID to the newer version and it retains the old version.

      The newest version of any object in a version-enabled bucket is known as the current version of that object.

      So in this case, the object called Winky.JPEG with an ID of 2222222, which is actually a dog picture, that is the current version of this object.

      Now if an object is accessed without explicitly indicating to S3 which version is required, then it's always the current version which will be returned.

      But you've always got the ability of requesting an object from S3 and providing the ID of a specific version to get that particular version back rather than the current version.

      So versions can be individually accessed by specifying the ID, and if you don't specify the ID, then it's assumed that you want to interact with the current version, the most recent version.

      Now versioning also impacts deletions.

      Let's say we've got these two different versions of Winky.JPEG stored in a version-enabled bucket.

      If we indicate to S3 that we want to delete the object and we don't give any specific version ID, then what S3 will do is try a new special version of that object known as a delete marker.

      Now the delete marker essentially is just a new version of that object, so S3 doesn't actually delete anything, but the delete marker makes it look deleted.

      In reality though, it's just hidden.

      The delete marker is a special version of an object which hides all previous versions of that object.

      But you can delete the delete marker which essentially undeletes the object, returning the current version to being active again, and all the previous versions of the object still exist, accessible using their unique version ID.

      Now even with versioning enabled, you can actually fully delete a version of an object, and that actually really deletes it.

      To do that, you just need to delete an object and specify the particular version ID that you want to remove.

      And if you are deleting a particular version of an object and the version that you're deleting is the most recent version, so the current version, then the next most recent version of that object then becomes the current version.

      Now some really important points that you need to be aware about object versioning.

      I've mentioned this at the start of the lesson, it cannot be switched off, it can only be suspended.

      Now why that matters is that when versioning is enabled on a bucket, all the versions of that object stay in that bucket, and so you're consuming space for all of the different versions of an object.

      If you have one single object that's 5 gig in size, and you have five versions of that object, then that's 5 times 5 gig of space that you're consuming for that one single object, and it's multiple versions.

      And logically, you'll build for all of those versions of all of those objects inside an S3 bucket, and the only way that you can zero those costs out is to delete the bucket and then re-upload all those objects to a bucket without versioning enabled.

      That's why it's important that you can't disable versioning.

      You can only suspend it, and when you suspend it, it doesn't actually remove any of those old versions, so you're still built for them.

      Now there's one other relevant feature of S3 which does make it to the exam all the time, and that's known as MFA delete.

      Now MFA delete is something that's enabled within the versioning configuration on a bucket.

      And when you enable MFA delete, it means that MFA is required to change bucket versioning state.

      So if you move from enable to suspend it or vice versa, you need this MFA to be able to do that, and also MFA is required to delete any versions of an object.

      So to fully delete any versions, you need this MFA token.

      Now the way that this works is that when you're performing API calls in order to change a bucket to versioning state or delete a particular version of an object, you need to provide the serial number of your MFA token as well as the code that it generates.

      You concatenate both of those together, and you pass that along with any API calls to interact how you delete versions or change the versioning state of a bucket.

      Okay, so that's all of the theory for object versioning inside S3.

      And at this point, that's everything I wanted to cover in this license.

      I'll go ahead and complete the video, and when you're ready, I'll look forward to you joining me in the next.

    1. Welcome back in this demo lesson you're going to gain some practical experience of working with the versioning feature of S3.

      So to get started just make sure that you're logged in to the management account of the organization, so the general account, and then make sure that you've got the Northern Virginia region selected, so US-EAS-1.

      Now there is a link attached to this lesson which you need to click on and then extract.

      This is going to contain all of the files that you'll be using throughout this demo.

      So go ahead and click on that link, extract it, and it should create a folder called S3_Versioning.

      Once you've confirmed that you're logged in and have the right region selected, then go ahead and move to the S3 console.

      So you can get there either using the recently visited services, or you can type S3 into the Find Services box and click to move to the S3 console.

      Now to demonstrate versioning we're going to go ahead and create an S3 bucket, we're going to set it up for static website hosting, enable versioning, and then experiment with some objects and just observe how versioning changes the default behavior inside an S3 bucket.

      So go ahead and click on Create bucket.

      As long as the bucket name is unique, its specific name isn't important because we won't be using it with Route 53.

      So just give the bucket a name and make sure that it's something unique.

      So I'm going to use AC_Bucket_13337.

      You should pick something different than me and different from something that any other student would use.

      Once you've selected a unique bucket name, just scroll down and uncheck Block All Public Access.

      We're going to be using this as a static website hosting bucket, so this is fine.

      And we'll need to acknowledge that we understand the changes that we're making, so check this box, scroll down a little bit more, and then under bucket versioning we're going to click to enable versioning.

      Keep scrolling down and at the bottom click on Create bucket.

      Next, go inside the bucket, click on Properties, scroll all the way down to the bottom, and we need to enable static website hosting.

      So click on Edit, check the box to enable static website hosting.

      For hosting type, we'll set it to host a static website, and then for the index document, just type index.html, and then for the error document, type error.html.

      Once you've set both of those, you can scroll down to the bottom and click on Save Changes.

      Now as you learned in the previous demo lesson, just enabling static website hosting isn't enough to allow access, we need to apply a bucket policy.

      So click on the permissions tab, scroll down, and under bucket policy click on Edit.

      Now inside the link attached to this lesson, which you should have downloaded and extracted, there should be a file called bucket_policy.json, which is an example bucket policy.

      So go ahead and open that file and copy the contents into your clipboard, move back to the console and paste it into the policy box, and we need to replace this example bucket placeholder with the ARN for this bucket.

      So copy the bucket ARN into your clipboard by clicking this icon.

      Because this ARN references objects in this bucket, and we know this because it's got forward slash star at the end, we need to replace only the first part of this placeholder ARN with the actual bucket ARN from the top.

      So select from the A all the way up to the T, so not including the forward slash and the star, and then paste in the bucket ARN that you copied onto your clipboard.

      Once you've done that, you can scroll down and then click on Save Changes.

      Next, click on the objects tab, and we're going to upload some of the files that you downloaded from the link attached to this lesson.

      So click on Upload, and first we're going to add the files.

      So click on Add Files, then you'll need to go to the location where you downloaded and extracted the file that's attached to this lesson.

      And once you're there, go into the folder called S3_Versioning, and you'll see a folder called Website.

      Open that folder, select index.html and click on Open, and then click on Add Folder, and select the IMG folder that's also in that same location.

      So select that folder and then click on Upload.

      So this is going to upload an index.html object, and it's going to upload a folder called IMG which contains winky.jpeg.

      Once you've done that, scroll down to the bottom and just click on Upload.

      Now once the upload's completed, you can go ahead and click on Close, and what you'll see in the Objects dialog inside the bucket is index.html and then a folder called IMG.

      And as we know by now, S3 doesn't actually have folders it uses prefixes, but if we go inside there, you'll see a single object called winky.jpeg.

      Now go back to the bucket, and what we're going to do is click on Properties, scroll down to the bottom, and then click on this icon to open our bucket in a new browser tab.

      All being well, you should see AnimalsForLife.org, Animal of the Week, and a picture of my one-eyed cat called winky.

      So this is using the same architecture as the previous demo lesson where you experienced static website hosting.

      What we're going to do now though is experiment with versions.

      So go back to the main S3 console, scroll to the top, and click on Objects.

      So because we've got versioning enabled on this bucket, as I talked about in the previous theory lesson, it means that every time you upload an object to this S3 bucket, it's assigned a version ID.

      And if you upload an object with the same name, then instead of overwriting that object, it just creates a new version of that object.

      Now with versioning enabled and using the default settings, we don't see all the individual versions, but we can elect to see them by toggling this Show Versions toggle.

      So go ahead and do that.

      Now you'll see that every object inside the S3 bucket, you'll see a particular version ID, and this is a unique code which represents this particular version of this particular object.

      So if we go inside the IMG folder, you'll see that we have the same for winkeep.jpeg.

      Toggle Show Versions to Disable, and you'll see that that version ID disappears.

      What I want you to do now is to click on the Upload button inside this IMG folder.

      So click on Upload, and then click on Add Files.

      Now inside this Lessons folder, so S3 versioning, at the top level you've got a number of folders.

      You have Website, which is what you uploaded to this S3 bucket, and this Image folder contains winkeep.jpeg.

      So this is a particular file, winkeep.jpeg, that contains the picture of winkeep my one-eyed cat.

      Now if you expand version 1 and version 2, you might be able to tell that version 1 is the same one-eyed cat, and we can expand that and say that it is actually winkeep.

      Inside version 2 we have an object with the same name, but if we expand this, this is not winkeep, this is a picture of truffles.

      So let's say that an administrator of this bucket makes a mistake and uploads this second version of winkeep.jpeg, which is not actually winkeep, it's actually truffles the cat.

      But let's say that we do this, so we select winkeep.jpeg from the version 2 folder, and we click on Open.

      Once we've selected that for upload, we scroll all the way down to the bottom and click on Upload.

      That might take a few seconds to complete the upload because these are relatively large image files, but once it's uploaded you can click on Close.

      So now we're still inside this image folder, and if we refresh, all we can see is one object, winkeep.jpeg.

      So it looks with this default configuration of the user interface, like we've overwritten a previous object with this new object.

      And if we go back to the tab which has got the static website open and hit refresh, you'll see that this image has indeed been replaced by the truffles image.

      So even though it's called winkeep.jpeg, this is clearly truffles.

      Now if we go back to the S3 console, and now if we enable the versions toggle, now we can see that we've got two different versions of this same object.

      We've got the original version at the bottom and a new version at the top.

      And note how both of these have different version IDs.

      Now what S3 does is it always picks the latest version whenever you use any operations which simply request that one object.

      So if we just request the object like we're doing with the static website hosting, then it will always pick the current or the latest version of this object.

      But we do still have access to the older versions because we have versioning enabled on this bucket.

      Nothing is ever truly deleted as long as we're operating with objects.

      So let's experiment with exactly what functionality this gives us.

      Go ahead and toggle show versions.

      Once you've done that, select the winkeep.jpeg object and then click delete.

      You'll need to type or copy and paste delete into this delete objects box and then click on delete.

      Before we do that, note what it says at the top.

      Deleting the specified objects adds delete markers to them.

      If you need to undo the delete action, you can delete the delete markers.

      So let's explore what this means.

      Go ahead and click on delete objects.

      And once it's completed, click on close.

      Now how this looks at the moment, we're still in the image folder.

      And because we've got show version set to off, it looks like we deleted the object.

      But this is not what's occurred because we've got versioning enabled.

      What's actually occurred is this is added a new version of this object.

      But instead of an actual new version of the object, it's simply added a delete marker as that new version.

      So if we toggle show versions back to on, now what we see are the previous versions of winkeep.jpeg.

      So the original version at the bottom and the one that we replaced in the middle.

      And then at the top we have this delete marker.

      Now the delete marker is the thing which makes it look to be deleted in the console UI when we have show version set to off.

      So this is how S3 handles deletions when versioning is enabled.

      If you're interacting with an object and you delete that object, it doesn't actually delete the object.

      It simply adds a delete marker as the most recent version of that object.

      Now if we just select that delete marker and then click on delete, that has the effect of undeleting the object.

      Now it's important to highlight that because we're dealing with object versions, anything that we do is permanent.

      If you're operating with an object and you have versioning enabled on a bucket, if you overwrite it or delete it, all it's going to do is either add a new version or add a delete marker.

      When you're operating with versions, everything is permanent.

      So in this case we're going to be permanently deleting the delete marker.

      So you need to confirm that by the typing or copying and pasting permanently delete into this box and click on delete objects.

      What this is going to do is delete the delete marker.

      So if we click on close, now we're left with these two versions of winkeep.jpeg so we've deleted the delete marker.

      If we toggle show versions to off, we can see that we now have our object back in the bucket.

      If we go back to static website hosting and refresh, we can see though that it's still truffle.

      So this is a mistake.

      It's not actually winky in this particular image.

      So what we can do is go back to the S3 console, we can enable show versions.

      We know that the most recent version is actually truffles rather than winky.

      So what we can do is select this incorrect version, so the most recent version and select delete.

      Now again, we're working with an object version.

      So this is permanent.

      You need to make sure that this is what you intend.

      In our case it is.

      So you need to either type or copy and paste permanently delete into the box and click on delete objects.

      Now this is going to delete the most recent version of this object.

      What happens when you do that is it makes the next most recent version of that object the current or latest version.

      So now this is the original version of winky.jpeg, the one that we first uploaded to this bucket.

      So this is now the only version of this object.

      If we go back to the static website hosting tab and hit refresh, this time it loads the correct version of this image.

      So this is actually winky my one-eyed cat.

      So this is how you can interact with versioning in an S3 bucket.

      Whenever it's enabled, it means that whenever you upload an object to the same name instead of overwriting, it simply creates a new version.

      Whenever you delete an object, it simply adds a delete marker.

      When you're operating with objects, it's always creating new versions or adding delete markers.

      But when you're working with particular versions rather than objects, any operations are permanent.

      So you can actually delete specific versions of an object permanently and you can delete delete markers to undelete that object.

      Now it's not possible to turn off versioning on a bucket.

      Once it's enabled on that bucket, you don't have the ability to disable it.

      You only have the ability to suspend it.

      Now when you suspend it, it stops new versions being created, but it does nothing about the existing versions.

      The only way to remove the additional costs for a version-enabled bucket is either to delete the bucket and then reload the objects to a new bucket, or go through the existing bucket and then manually purge any specific versions of objects which aren't required.

      So you need to be careful when you're enabling versioning on a bucket because it can cause additional costs.

      If you have a bucket where you're uploading objects over and over again, specifically of their large objects, then if you have versioning enabled, you can incur significantly higher costs than if you have a bucket which doesn't have a versioning enabled.

      So that's something you need to keep in mind.

      If you enable versioning, you need to manage those versions of those objects inside the bucket.

      With that being said, let's tidy up.

      So let's go back to the main S3 console, select the bucket, click on Empty, copy and paste or type "Permanently Delete" and click on Empty.

      When it's finished, click on Exit, and with the bucket still selected, click on Delete.

      Copy and paste or type the name of the bucket and confirm it with the delete bucket.

      I want you to build out the accounties back in the same state as it was at the start of this demo lesson.

      Now at this point, that's everything that I want you to do in this demo lesson.

      You've gained some practical exposure with how to deal with object versions inside an S3 bucket.

      At this point, go ahead and complete this video, and when you're ready, I'll afford you joining me in the next lesson.

    1. Welcome back.

      And in this demo lesson, you're going to get some experience using the S3 static website hosting feature, which I talked about in the previous lesson.

      Now, to get started, just make sure that you're logged in to the management account of the organization, and that you're using the IAM Admin user.

      So this just makes sure that you have admin permissions over the general or management account of the organization.

      Also, make sure that you have the Northern Virginia region selected, which is US-EAST-1.

      Normally with S3, when you're interacting with the product, you're doing so using the AWS console UI or the S3 APIs.

      And in this demo lesson, you'll be enabling a feature which allows S3 to essentially operate as a web server.

      It allows anybody with a web browser to interact with an S3 bucket, load an index page, and load pictures or other media that are contained within that bucket using standard HTTP.

      So that's what we're going to do.

      And to get started, we need to move across to the S3 console.

      So either use S3 in recently visited services, or you can click on the services dropdown, type S3, and then click it in this list.

      Now that we're at the S3 console, we're going to create an S3 bucket.

      Now, if you chose to register a domain earlier in the course, like I did, so I registered animalsforlife.io, then we're going to connect this S3 bucket with the custom domain that we registered so we can access it using that domain.

      If you chose not to use a domain, then don't worry, you can still do this demo.

      What you need to do is to go ahead and click on Create a bucket.

      Now for the bucket name, if you are not using a custom domain, then you can enter whatever you want in this bucket name as long as it's unique.

      If you did register a custom domain and you want to use this bucket with that domain, then you need to enter a DNS formatted bucket name.

      So in my case, I'm going to create a bucket which is called Top 10.

      It's going to store the world's best cappages, the Top 10 cappages in the world ever.

      And it's going to be part of the animalsforlife.io domain.

      And so at the end of this, I'm going to add dot and then animalsforlife.io.

      And if you've registered your own custom domain, then obviously you need to add your own domain at the end.

      You can't use the same name as me.

      Once you've entered that name and just scroll down and uncheck Block All Public Access, this is a safety feature of S3.

      But because we're intentionally creating an S3 bucket to be used as a static website, we need to uncheck this box.

      Now, unchecking this box means that you will be able to grant public access.

      It doesn't mean that public access is granted automatically when you uncheck this box.

      They're separate steps.

      You will, though, need to acknowledge that you understand the risks of unticking that box.

      So check this box just to confirm that you understand.

      We'll be carefully configuring the security so you don't have to worry about any of those risks.

      And once you've set that, we can leave everything else as default.

      So just scroll all the way down to the bottom and click on Create Bucket.

      So the bucket's been created, but right now, this only allows access using the S3 APIs or the console UI.

      So we need to enable static website hosting.

      Now, to do that, we're going to click on the bucket.

      Once we have the bucket open, we're going to select the Properties tab.

      On the Properties tab, scroll all the way down to the bottom.

      And right at the very bottom, we've got static website hosting.

      And you need to click on the Edit button next to that.

      It's a simple yes or no choice at this point.

      So check the box to enable static website hosting.

      There are a number of different types of hosting.

      You can either just host a static website, which is what we'll choose, or you can redirect requests for an object.

      So this allows you to redirect to a different S3 bucket.

      We'll be covering this later in the course for now.

      Just leave this selected.

      So host a static website.

      Now, in order to use the static website hosting feature, you'll need to provide S3 with two different documents.

      The index document is used as the home or default page for the static website hosting.

      So if you don't specify a particular object when you're browsing to the bucket, for example, winky.jpg, if you just browse to the bucket itself, then the index document is used.

      And we're going to specify index.html.

      So this means that an object called index.html will be loaded if we don't specify one.

      Now, the error document is used whenever you have any errors.

      So if you specify that you want to retrieve an object from the bucket, which doesn't exist, the error document is used.

      And for the error document, we're going to call this error.html.

      So these two values always need to be provided when you enable static website hosting.

      So now we've provided those, we can scroll down and click on save changes.

      Now that that feature's enabled, if we just scroll all the way down to the bottom, you'll see that we have a URL for this bucket.

      So go ahead and copy that into your clipboard.

      We're going to need this shortly.

      So this is the URL that you'll use by default to browse to this bucket.

      Now, next, what we need to do is to upload some objects to the bucket, which this static website hosting feature is going to use.

      Now, to do that, scroll all the way to the top and just click on objects and then click on upload.

      So this is the most recent UI version for S3.

      And so you have the ability to add files or add folders.

      Now, we're going to use both of these.

      We're going to use the add files button to add the index.html and the error.html.

      And we're going to use the add folder to add a folder of images.

      So first, let's do the add files.

      So click on add files.

      Now, attached to this video is a link which downloads all of the assets that you'll need for this demo.

      So go ahead and click on that link to download the zip file and then extract that zip file to your local machine.

      And you'll need to move to the folder that you extracted from the zip file.

      It should be called static_website_hosting.

      So go to that folder.

      And then again, there should be a folder in there called website_files.

      So go ahead and click on there to go into that folder.

      Now, there are three things inside this folder, index.html, error.html and img.

      So we'll start by uploading both of these HTML documents.

      So select index.html and error.html and then click on open.

      And that will add both of these to this upload table.

      Next, click on add folder and then select the img folder and click on upload.

      So this has prepared all of these different objects ready to upload to this S3 bucket.

      If we scroll down, we'll see that the destination for these uploads is our S3 bucket and your name here will be different as long as it's the same as the name you picked for the bucket, that's fine.

      Go all the way to the bottom and then go ahead and click on upload.

      And that will upload the index.html, the error.html and then the folder called img as well as the contents of that folder.

      So at this point, that's all of the objects uploaded to the S3 bucket and we can go ahead and click on close.

      So now let's try browsing to this bucket using static website hosting.

      So go ahead and click on properties, scroll all the way down to the bottom and here we've got the URL for this S3 bucket.

      So go ahead and copy this into your clipboard, open a new tab and then open this URL or click on this symbol to open it in a new tab.

      What you'll see is a 403 forbidden error and this is an access denied.

      You're getting this error because you don't have any permissions to access the objects within this S3 bucket.

      Remember, S3 is private by default and just because we've enabled static website hosting doesn't mean that we have any permissions to access the objects within this S3 bucket.

      We're accessing this bucket as an anonymous or unauthenticated user.

      So we have no method of providing any credentials to S3 when we're accessing objects via static website hosting.

      So we need to give permissions to any unauthenticated or anonymous users to access the objects within this bucket.

      So that's the next thing we need to do.

      We need to grant permissions to be able to read these objects to any unauthenticated user.

      So how do we do that?

      The third method is to use a bucket policy.

      So that's what I'm gonna demonstrate in order to grant access to these objects.

      Now to add a bucket policy, we need to select the permissions tab.

      So click on permissions and then below block public access, there's a box to specify a bucket policy.

      So click on edit and we need to add a bucket policy.

      Now also in the folder that you extracted from this lessons zip file is a file called bucket_policy.json and this is a generic bucket policy.

      So this bucket policy has an effect of allow and it applies to any principle because we have this star wild card and because the effect is allow, it grants any principle, the ability to use the S3 get object action which allows anyone to read an object inside an S3 bucket and it applies to this resource.

      So this is a generic template, we need to update it, but go ahead and copy it into your clipboard, go back to the S3 console and paste it into this box.

      Now we need to replace this generic ARN, so this example bucket ARN.

      So what I want you to do is to copy this bucket ARN at the top of the screen.

      So copy this into your clipboard and we need to replace part of this template ARN with what we've just copied.

      Now an important point to highlight is that this ARN has forward slash star on the end because this ARN refers to any objects within this S3 bucket.

      So we need to select only the part before the forward slash.

      So starting at the A and then ending at the end of example bucket and then just paste in the ARN at our bucket that we just copied into our clipboard.

      What you should end up with is this full ARN with the name of the bucket that you created and then forward slash star.

      And once you've got that, go ahead and click on save changes.

      This applies a bucket policy which allows any principle, so even unauthenticated principles, the ability to get any of the objects inside this bucket.

      So this means that any principle will be able to read objects inside this bucket.

      At this point, assuming everything's okay, if you've still got the tab open to the bucket, then go back to that tab and hit refresh.

      And what you should see is the top 10 animals in the world.

      So position number one, we've got Merlin.

      At position number two, we've got Merlin again.

      Position number three, another Merlin.

      Four, still Merlin.

      And then Merlin again at number five.

      At number six, we've got Boris.

      So the token non Merlin cat.

      Number seven, Samson, another token non Merlin cat.

      And then number eight, we've got different cat one.

      He looks quite a lot like Merlin.

      Number nine, different cat two, again, kind of looks like Merlin.

      And then number 10, we've got the family.

      And then you might not have guessed this, but this entire top 10 contest was judged by, you guessed it, Merlin.

      So what you're loading here is the index.html document inside the bucket.

      So we haven't specified an object to load.

      And because of that, it's using the index document that we specified on the bucket.

      We can load the same object by typing specifically index.html on the end, and that will load in the same object.

      Now, if we specify an object which doesn't exist, so let's say we used wrong index.html, then instead of the index document, now it's going to load the error document.

      So this is the error document that you specified, which is loading error.html.

      So this is just an example of how you can configure an S3 bucket to act as a standard static website.

      So what it's doing is loading in the index.html object inside the bucket.

      And that index.html is loading in images, which are also stored in the bucket.

      So if I right click and copy the image location and open this in a new tab, this is essentially just loading this image from the same S3 bucket.

      So it's loading it from this folder called img, and it's called Merlin.jpeg.

      It's just an object loading from within the bucket.

      Now if I go back to the S3 console and just move across to the properties tab and then scroll down, so far in this lesson, you've been accessing this bucket using the bucket website endpoint.

      So this is an endpoint that's derived from the name of the bucket.

      Now your URL will be different because you will have called your bucket name something else.

      Now if you chose to register a custom domain name at the start of this course, you can customize this further.

      As long as you call the bucket the same as the DNS name that you want to use, you can actually use Route 53 to assign a custom DNS name for this bucket.

      So this part of the demo you'll only be able to do if you've registered a domain within Route 53.

      If you haven't, you can skip to the end of this demo where we're going to tidy up.

      But if you want to customize this using Route 53, then you can click on the services dropdown and type Route 53 and then click to move to the Route 53 console.

      Once you're there, you can click on hosted zones and you should have a hosted zone that matches the domain that you registered at the start of the course.

      Go inside that and click on create record.

      Now we're going to be creating a simple routing record.

      So make sure that's selected and then click on next.

      And we're going to define a simple record.

      Now I'm going to type the first part of the name of the bucket.

      So I used top10.animalsforlive.io is my bucket name.

      So I'm going to put top 10 in this box.

      Now, because we want to point this at our S3 bucket, we need to choose an endpoint in this dropdown.

      So click in this dropdown and then scroll down and we're going to pick alias to S3 website endpoint.

      So select that.

      Next, you need to choose the region and you should have created the S3 bucket in the US East 1 region because this is the default for everything that we do in the course.

      So go ahead and type US-EAS-1 and then select US East Northern Virginia and you should be able to click in enter S3 endpoint and select your bucket name.

      Now, if you don't see your bucket here, then either you've picked the wrong region or you've not used the same name in this part of the record name as you picked for your bucket.

      So make sure this entire name, so this component plus the domain that you use matches the name that you've selected for the bucket.

      Assuming it does, you should be able to pick your bucket in this dropdown.

      Once you've selected it, go ahead and click on define simple record.

      And once that's populated in the box, click on create records.

      Now, once this record's created, you might have to wait a few moments, but you should find that you can then open this bucket using this full DNS name.

      So there we go.

      It opens up the same bucket.

      So we've used Route 53 and we've integrated it using an alias to our S3 endpoint.

      Now, again, you can only do this if you create a bucket with the same name as the fully qualified domain name that we just configured.

      So this is an example of a fully qualified domain name.

      Now, this is the host component of DNS and this is the domain component.

      So together they make up a fully qualified domain name and for this to work, you need to create an S3 bucket with the same bucket name as this fully qualified domain name.

      And that's what I did at the start of this lesson, which is why it works for me.

      And as long as you've done the same, as long as you've registered a custom domain, as long as you've called the bucket the same as what you're creating within Route 53, then you should be able to reference that bucket and then access it using this custom URL.

      At this point, we're going to tidy up.

      So go back to the Route 53 console and select this record that you've created and then click on delete.

      You'll need to confirm it by clicking delete again.

      Then we need to go back to the S3 console, select the bucket that you've created, click on empty, and you'll need to either type or copy and paste, permanently delete into this box, and then click on empty.

      It'll take a few minutes to empty the bucket.

      Once it's completed, click on exit.

      And with the bucket still selected, click on delete to delete the bucket.

      And you'll need to confirm that by either typing or copy and pasting the name of the bucket and then click delete bucket.

      Now, at this point, that's everything that you need to do in this lesson.

      It's just an opportunity to experience the theory that you learned in the previous lesson.

      Now, there's a lot more that you can do with static website hosting and I'll be going into many more complex examples later on in the course.

      But for now, this is everything that you need to do.

      So go ahead and complete this video.

      And when you're ready, I'll look forward to you joining me in the next.

    1. When philosophers talk about common sense, they mean specific claims based on direct sense perception, which are true in a relatively fundamental sense.

      I think that in our day in age today sometimes we take the phrase it's just common sense a little to far. We never know what certain peoples common sense is, sometimes those words can be hurtful to someone that may not be understanding what is being talked about

    1. Welcome back and in this lesson I want to start talking about S3 security in more detail. Starting with bucket policies which are a type of AWS resource policy. So by now you know the drill, let's jump in and get started.

      Now before we start I want to repeat one thing and you have heard me say this before, but I'm going to say it again over and over. S3 is private by default. Everything that we can do to control S3 permissions is based on this starting point. The only identity which has any initial access to an S3 bucket is the account root user of the account which owns that bucket, so the account which created it. Anything else, so any other permissions have to be explicitly granted. And there are a few ways that this can be done.

      The first way is using an S3 bucket policy. And an S3 bucket policy is a type of resource policy. A resource policy is just like an identity policy, but as the name suggests, they're attached to resources instead of identities, in this case an S3 bucket. Resource policies provide a resource perspective on permissions. The difference between resource policies and identity policies is all about this perspective. With identity policies you're controlling what that identity can access. With resource policies you're controlling who can access that resource. So it's from an inverse perspective. One is identities and one is resources.

      Now identity policies have one pretty significant limitation. You can only attach identity policies to identities in your own account. And so identity policies can only control security inside your account. With identity policies you have no way of giving an identity in another account access to an S3 bucket. That would require an action inside that other account. Resource policies allow this. They can allow access from the same account or different accounts because the policy is attached to the resource and it can reference any other identities inside that policy. So by attaching the policy to the resource and then having flexibility to be able to reference any other identity, whether they're in the same account or different accounts, resource policies therefore provide a great way of controlling access for a particular resource, no matter what the source of that access is.

      Now think about that for a minute because that's a major benefit of resource policies, the ability to grant other accounts access to resources inside your account. They also have another benefit, resource policies can allow or deny anonymous principals. Identity policies by design have to be attached to a valid identity in AWS. You can't have one attached to nothing. Resource policies can be used to open a bucket to the world by referencing all principals, even those not authenticated by AWS. So that's anonymous principals. So bucket policies can be used to grant anonymous access.

      So two of the very common uses for bucket policies are to grant access to other AWS accounts and anonymous access to a bucket. Let's take a look at a simple visual example of a bucket policy because I think it will help you understand how everything fits together. There's a demo lesson coming up soon where you'll implement one as part of the mini project. So you will get some experience soon enough of how to use bucket policies.

      Let's say that we have an AWS account and inside this account is a bucket called Secret Cat Project. Now I can't say what's inside this bucket because it's a secret, but I'm sure that you can guess. Now attached to this bucket is a bucket policy. Resource policies have one major difference to identity policies and that's the presence of an explicit principal component. The principal part of a resource policy defines which principals are affected by the policy. So the policy is attached to a bucket in this case, but we need a way to say who is impacted by the configuration of that policy. Because a bucket policy can contain multiple statements, there might be one statement which affects your account and one which affects another account, as well as one which affects a specific user, the principal part of a policy or more specifically the principal part of a statement in a policy defines who that statement applies to, which identity is which principals.

      Now in an identity policy this generally isn't there because it's implied that the identity which the policy is applied to is the principal. That's logical right? Your identity policy by definition applies to you so you are the principal. So a good way of identifying if a policy is a resource policy or an identity policy is the presence of this principal component. If it's there it's probably a resource policy. In this case the principal is a wild card, a star, which means any principal. So this policy applies to anyone accessing the S3 bucket.

      So let's interpret this policy. Well first the effect is allow and the principal is star, so any principal. So this effect allows any principal to perform the action S3 get object on any object inside the secret cat project S3 bucket. So in effect it allows anyone to read any objects inside this bucket. So this would equally apply to identities in the same AWS account as the bucket. It could also apply to other AWS accounts, or partner account. And crucially it also applies to anonymous principals. So principals who haven't authenticated to AWS. Bucket policies should be your default thought when it comes to granting anonymous access to objects in buckets and they're one way of granting external accounts that same access. They can also be used to set the default permissions on a bucket. If you want to grant everyone access to Boris's picture for example and then grant certain identities extra rights or even deny certain rights then you can do that. Bucket policies are really flexible. They can do many other things.

      So let's quickly just look at a couple of common examples. Bucket policies can be used to control who can access objects even allowing conditions which block specific IP addresses. In this example this bucket policy denies access to any objects in the secret cat project bucket unless your IP address is 1.3.3.7. The condition block here means this statement only applies if this condition is true. So if your IP address, the source IP address is not 1.3.3.7, then the statement applies and access is denied. If your IP address is 1.3.3.7, then this condition is not met because it's a not IP address condition. So if your IP address is this IP address, the condition is not matched and you get any other access that's applicable. Essentially this statement, which is a deny, does not apply.

      Now, bucket policies can be much more complex. In this example, one specific prefix in the bucket, remember this is what a folder really is inside a bucket, so one specific prefix called Boris is protected with MFA. It means that accesses to the Boris folder in the bucket are denied if the identity that you're using does not use MFA. The second statement allows read access to objects in the whole bucket. Because an explicit deny overrides an allow, the top statement applies to just that specific prefix in the bucket, so just Boris. Now, I won't labour on about bucket policies because we'll be using them a fair bit throughout the course, but they can range from simple to complex. I will include a link in the lesson description with some additional examples that you can take a look through if you're interested.

      In summary though, a resource policy is associated with a resource. A bucket policy, which is a type of resource policy, is logically associated with a bucket, which is a type of resource. Now, there can only be one bucket policy on a bucket, but it can have multiple statements. If an identity inside one AWS account is accessing a bucket, also in that same account, then the effective access is a combination of all of the applicable identity policies plus the resource policy, so the bucket policy. For any anonymous access, so access by an anonymous principal, then only the bucket policy applies, because logically, if it's an anonymous principal, it's not authenticated and so no identity policies apply.

      Now, if an identity in an external AWS account attempts to access a bucket in your account, your bucket policy applies as well as anything that's in their identity policies. So there's a two-step process if you're doing cross-account access. The identity in their account needs to be able to access S3 in general and your bucket, and then your bucket policy needs to allow access from that identity, so from that external account.

      Now, there is another form of S3 security. It's used less often these days, but I wanted to cover it anyway. Access control lists or ACLs are ways to apply security to objects or buckets. There is a sub-resource of that object or of that bucket. Remember in the S3 introduction lesson earlier in the course, I talked about sub-resources. Well, this is one of those sub-resources. Now, I almost didn't want to talk about ACLs because they are legacy. AWS don't even recommend their use and prefer that you use bucket policies or identity policies. But as a bare minimum, I want you to be aware of their existence.

      Now, part of the reason that they aren't used all that often and that bucket policies have replaced much of what they do is that they're actually inflexible and only allow very simple permissions. They can't have conditions like bucket policies and so you're restricted to some very broad conditions. Let me show you what I mean. This is an example of what permissions can be controlled using an ACL. Now, apologies for the wall of text, but I think it's useful to visualize it all at once. There are five permissions which can be granted in an ACL. Read, write, readACP, writeACP and full control. That's it. So it's already significantly less flexible than an identity or a resource policy. What these five things do depend on if they're applied to a bucket or an object. Read permissions, for example, on a bucket allow you to list all objects in that bucket, whereas write permissions on a bucket allow the grantee, which is the principal being granted those permissions, the ability to overwrite and delete any object in that bucket. Read permissions on an object allow the grantee just to read the object specifically as well as its metadata.

      Now, with ACLs you either configure an ACL on the bucket, or you configure the ACL on an object. But you don't have the flexibility of being able to have a single ACL that affects a group of objects. You can't do that. That's one of the reasons that a bucket policy is significantly more flexible. It is honestly so much less flexible than a bucket policy to the extent where I won't waste your time with it anymore. It's legacy, and I suspect at some point it won't be used anymore. If there are any specific places in the course which do require knowledge of ACLs, I'll mention it. Otherwise, it's best to almost ignore the fact that they exist.

      Now, before we finish up, one final feature of S3 permissions, and that's the block public access settings. In the overall lifetime of the S3 product, this was actually added fairly recently, and it was added in response to lots of public PR disasters where buckets were being configured incorrectly and being set so that they were open to the world. This resulted in a lot of data leaks, and the root cause was a mixture of genuine mistakes or administrators who didn't fully understand the S3 permissions model.

      So consider this example, an S3 bucket with resource permissions granting public access. Until block public access was introduced, if you had public access configured, the public could logically access a bucket. Public access in this sense is read only to any objects defined in a resource policy on a bucket, so there's no restrictions. Public access is public access. Block public access added a further level of security, another boundary. And on this boundary is the block public access settings, which apply no matter what the bucket policies say, but they apply to just the public access, so not any of the defined AWS identities. So these settings will only apply to an anonymous principal, somebody who isn't an AWS identity, attempt to access a bucket using these public access configurations.

      Now these settings can be set when you create the bucket and adjusted afterwards. They're pretty simple to understand. You can choose the top option which blocks any public access to the bucket, no matter what the resource policy says. It's a full override, a failsafe. Or you can choose the second option which allows any public access granted by any existing ACLs when you enable the setting but it blocks any new ones. The third option blocks any public access granted by ACLs no matter if it was enabled before or after the block public access settings were enabled. The fourth setting allows any existing public access granted by bucket policies or access point policies so anything enabled at the time when you enable this specific block public access setting, they're allowed to continue but it blocks any new ones. The fifth option blocks both existing and new bucket policies from granting any public access.

      Now they're simple enough and they function as a final failsafe. If you're ever in a situation where you've granted some public access and it doesn't work, these are probably the settings which are causing that inconsistency. And don't worry, I'll show you where these are accessed in the demo lesson.

      Now before we finish up, just one final thing I want to cover and this is an exam at PowerUp. So these are just some key points on how to remember all of the theory that I've discussed in this lesson. When I first started in AWS, I found it hard to know from instinct when to use identity policies versus resource policies versus ACLs. Choosing between resource policies and identity policies much of the time is a preference thing. So do you want to control permissions from the perspective of a bucket or do you want to grant or deny access from the perspective of the identities accessing a bucket? Are you looking to configure one user accessing 10 different buckets or 100 users accessing the same bucket? It's often a personal choice. A choice on what makes sense for your situation and business. So there's often no right answer but there are some situations where one makes sense over the other.

      If you're granting or denying permissions on lots of different resources across an AWS account, then you need to use identity policies because not every service supports resource policies. And besides, you would need a resource policy for each service so that doesn't make sense if you're controlling lots of different resources. If you have a preference for managing permissions all in one place, that single place needs to be IAM, so identity policies would make sense. IAM is the only single place in AWS you can control permissions for everything. You can sometimes use resource policies but you can use IAM policies all the time. If you're only working with permissions within the same account so no external access, then identity policies within IAM are fine because with IAM you can only manage permissions for identities that you control in your account. So there are a wide range of situations where IAM makes sense and that's why most permissions control is done within IAM. But there are some situations which are different. You can use bucket policies or resource policies in general if you're managing permissions on a specific product. So in this case S3. If you want to grant a single permission to everybody accessing one resource or everybody in one account, then it's much more efficient to use resource policies to control that base level permission. If you want to directly allow anonymous identities or external identities from other AWS accounts to access a resource, then you should use resource policies.

      Now finally, and I know this might seem like I'm anti-access control list, which is true, but so are AWS, never use ACLs unless you really need to. And even then, consider if you can use something else. At this point in time, if you are using an ACL, you have to be pretty certain that you can't use anything else because they're legacy and their inflexible and AWS are actively recommending against their use. So keep that in mind.

      Okay, well that's all of the theory that I wanted to cover in this lesson. I know it's been a lot, but we do have to cover this detailed level of security because it's needed in the exam. And you'll be using it constantly throughout the rest of this section and the wider course. At this point, though, go ahead and complete this video. And when you're ready, you can join me in the next where I'm going to be talking about another exciting feature of S3.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews

      Reviewer #1 (Public Review):

      Summary:

      Federer et al. tested AAVs designed to target GABAergic cells and parvalbumin-expressing cells in marmoset V1. Several new results were obtained. First, AAV-h56D targeted GABAergic cells with >90% specificity, and this varied with serotype and layer. Second, AAV-PHP.eB.S5E2 targeted parvalbumin-expressing neurons with up to 98% specificity. Third, the immunohistochemical detection of GABA and PV was attenuated near viral injection sites.

      Strengths:

      Vormstein-Schneider et al. (2020) tested their AAV-S5E2 vector in marmosets by intravenous injection. The data presented in this manuscript are valuable in part because they show the transduction pattern produced by intraparenchymal injections, which are more conventional and efficient.

      Our manuscript additionally provides detailed information on the laminar specificity and coverage of these viral vectors, which was not investigated in the original studies.

      Weaknesses:

      The conclusions regarding the effects of serotype are based on data from single injection tracks in a single animal. I understand that ethical and financial constraints preclude high throughput testing, but these limitations do not change what can be inferred from the measurements. The text asserts that "...serotype 9 is a better choice when high specificity and coverage across all layers are required". The data presented are consistent with this idea but do not make a strong case for it.

      We are aware of the limitations of our results on the AAV-h56D. We agree with the Reviewer that a single injection per serotype does not allow us to make strong statements about differences between the 3 serotypes. Therefore, in the revised version of the manuscript we have tempered our claims about such differences and use more caution in the interpretation of these data (Results p. 6 and Discussion p.10). Despite this weakness, we feel that these data still demonstrate high efficiency and specificity across cortical layers of transgene expression in GABA cells using the h56D promoter, at least with two of the 3 AAV serotypes we tested. We feel that in itself this is sufficiently useful information for the primate community, worthy of being reported. Due to cost, time and ethical considerations related to the use of primates, we chose not to perform additional experiments to determine precise differences among serotypes. Thus, for example, while it is possible that had we replicated these experiments, serotype 7 could have proven equally efficient and specific as the other two serotypes, we felt answering this question did not warrant additional experiments in this precious species.

      A related criticism extends to the analysis of Injection volume on viral specificity. Some replication was performed here, but reliability across injections was not reported. My understanding is that individual ROIs were treated as independent observations. These are not biological replicates (arguably, neither are multiple injection tracks in a single animal, but they are certainly closer). Idiosyncrasies between animals or injections (e.g., if one injection happened to hit one layer more than another) could have substantial impacts on the measurements. It remains unclear which results regarding injection volume or serotype would hold up had a large number of injections been made into a large number of marmosets.

      For the AAV-S5E2, we made a total of 7 injections (at least 2 at each volume), all of which, irrespective of volume, resulted in high specificity and efficiency for PV interneurons. Our conclusion is that larger volumes are slightly less specific, but the differences are minimal and do not warrant additional injections. Additionally, we kept all the other parameters across animals constant (see new Supplementary Table 1), all of our injections involved all cortical layers, and the ROIs we selected for counts encompassed reporter protein expression across all layers. To provide a better sense of the reliability of the results across injections, in the revised version of the manuscript we now provide results for each of the AAV-S5E2 injection case separately in a new Supplementary Table 2. The results in this table indicate the results are indeed rather consistent across cases with slightly greater specificity for injection volumes in the range of 105-180 nl.

      Reviewer #2 (Public Review):

      This is a straightforward manuscript assessing the specificity and efficiency of transgene expression in marmoset primary visual cortex (V1), for 4 different AAV vectors known to target transgene expression to either inhibitory cortical neurons (3 serotypes of AAV-h56D-tdTomato) or parvalbumin (PV)+ inhibitory cortical neurons in mice. Vectors are injected into the marmoset cortex and then postmortem tissue is analyzed following antibody labeling against GABA and PV. It is reported that: "in marmoset V1 AAV-h56D induces transgene expression in GABAergic cells with up to 91-94% specificity and 80% efficiency, depending on viral serotype and cortical layer. AAV-PHP.eB-S5E2 induces transgene expression in PV cells across all cortical layers with up to 98% specificity and 86-90% efficiency."

      These claims are largely supported but slightly exaggerated relative to the actual values in the results presented. In particular, the overall efficiency for the best h56D vectors described in the results is: "Overall, across all layers, AAV9 and AAV1 showed significantly higher coverage (66.1{plus minus}3.9 and 64.9%{plus minus}3.7)". The highest coverage observed is just in middle layers and is also less than 80%: "(AAV9: 78.5%{plus minus}9.1; AAV1: 76.9%{plus minus}7.4)".

      In the abstract, we indeed summarize the overall data and round up the decimals, and state that these percentages are upper bound but that they vary by serotype and layer while in the Results we report the detailed counts with decimals. To clarify this, in the revised version of the Abstract we have changed 80% to 79% and emphasize even more clearly the dependence on serotype and layer. We have amended this sentence of the Abstract as follows: “We show that in marmoset V1 AAV-h56D induces transgene expression in GABAergic cells with up to 91-94% specificity and 79% efficiency, but this depends on viral serotype and cortical layer.”

      For the AAV-PHP.eB-S5E2 the efficiency reported in the abstract (“86-90%) is also slightly exaggerated relative to the results: “Overall, across all layers coverage ranged from 78%{plus minus}1.9 for injection volumes >300nl to 81.6%{plus minus}1.8 for injection volumes of 100nl.”

      Indeed, the numbers in the Abstract are upper bounds, for example efficiency in L4A/B with S5E2 reaches 90%. To further clarify this important point, in the revised abstract we now state ”AAV-PHP.eB-S5E2 induces transgene expression in PV cells across all cortical layers with up to 98% specificity and 86-90% efficiency, depending on layer”.

      These data will be useful to others who might be interested in targeting transgene expression in these cell types in monkeys. Suggestions for improvement are to include more details about the vectors injected and to delete some comments about results that are not documented based on vectors that are not described (see below).

      Major comments:

      Details provided about the AAV vectors used with the h56D enhancer are not sufficient to allow assessment of their potential utility relative to the results presented. All that is provided is: "The fourth animal received 3 injections, each of a different AAV serotype (1, 7, and 9) of the AAV-h56D-tdTomato (Mehta et al., 2019), obtained from the Zemelman laboratory (UT Austin)." At a minimum, it is necessary to provide the titers of each of the vectors. It would also be helpful to provide more information about viral preparation for both these vectors and the AAVPHP.eB-S5E2.tdTomato. Notably, what purification methods were used, and what specific methods were used to measure the titers?

      We thank the Reviewer for this comment. In the revised version of the manuscript, we now provide a new Supplementary Table 1 with titers and other information for each viral vector injection. We also provide information regarding viral preparation in a new sections in the Methods entitled “ Viral Preparation”  (p12).

      The first paragraph of the results includes brief anecdotal claims without any data to support them and without any details about the relevant vectors that would allow any data that might have been collected to be critically assessed. These statements should be deleted. Specifically, delete: “as well as 3 different kinds of PV-specific AAVs, specifically a mixture of AAV1-PaqR4-Flp and AAV1-h56D-mCherry-FRT (Mehta et al., 2019), an AAV1-PV1-ChR2-eYFP (donated by G. Horwitz, University of Washington),” and delete “Here we report results only from those vectors that were deemed to be most promising for use in primate cortex, based on infectivity and specificity. These were the 3 serotypes of the GABA-specific pAAV-h56D-tdTomato, and the PV-specific AAVPHP.eB-S5E2.tdTomato.” These tools might in fact be just as useful or even better than what is actually tested and reported here, but maybe the viral titer was too low to expect any expression.

      These data are indeed anecdotal, but we felt this could be useful information, potentially preventing other primate labs from wasting resources, animals and time, particularly, as some of these vectors have been reported to be selective and efficient in primate cortex, which we have not been able to confirm. We made several injections in several animals of those vectors that failed either to infect a sufficient number of cells or turned out to be poorly specific. Therefore, the negative results have been consistent in our hands. But we agree with the Reviewer that our negative results could have depended on factors such as titer. In the revised version of the manuscript, following the reviewer’s suggestion, we have deleted this information.

      Based on the description in the Methods it seems that no antibody labeling against TdTomato was used to amplify the detection of the transgenes expressed from the AAV vectors. It should be verified that this is the case - a statement could be added to the Methods.

      That is indeed the case. We used no immunohistochemistry to enhance the reporter proteins as this was unnecessary. The native/ non-amplified tdT signal was strong. This is now stated in the methods (p.12).

      Reviewer #3 (Public Review):

      Summary:

      Federer et al. describe the laminar profiles of GABA+ and of PV+ neurons in marmoset V1. They also report on the selectivity and efficiency of expression of a PV-selective enhancer (S5E2). Three further viruses were tested, with a view to characterizing the expression profiles of a GABA-selective enhancer (h56d), but these results are preliminary.

      Strengths:

      The derivation of cell-type specific enhancers is key for translating the types of circuit analyses that can be performed in mice - which rely on germline modifications for access to cell-type specific manipulation - in higher-order mammals. Federer et al. further validate the utility of S5E2 as a PV-selective enhancer in NHPs.

      Additionally, the authors characterize the laminar distribution pattern of GABA+ and PV+ cells in V1. This survey may prove valuable to researchers seeking to understand and manipulate the microcircuitry mediating the excitation-inhibition balance in this region of the marmoset brain.

      Weaknesses:

      Enhancer/promoter specificity and efficiency cannot be directly compared, because they were packaged in different serotypes of AAV.

      The three different serotypes of AAV expressing reporter under the h56D promoter were only tested once each, and all in the same animal. There are many variables that can contribute to the success (or failure) of a viral injection, so observations with an n=1 cannot be considered reliable.

      This is an important point that was also brough up by Reviewer 1, which we have addressed in our reply-to-Reviewer 1. For clarity and convenience, below we copy our response to Reviewer 1.

      “We are aware of the limitations of our results on the AAV-h56D. We agree with the Reviewer that a single injection per serotype does not allow us to make strong statements about differences between the 3 serotypes. Therefore, in the revised version of the manuscript we will temper our claims about such differences and use more caution in the interpretation of these data. Despite this weakness, we feel that these data still demonstrate high efficiency and specificity across cortical layers of transgene expression in GABA cells using the h56D promoter, at least with two of the 3 AAV serotypes we tested. We feel that in itself this is sufficiently useful information for the primate community, worthy of being reported. Due to cost, time and ethical considerations related to the use of primates, we chose not to perform additional experiments to determine precise differences among serotypes. Thus, for example, while it is possible that had we replicated these experiments, serotype 7 would have proven equally efficient and specific as the other two serotypes, we felt answering this question did not warrant additional experiments in this precious species.”

      The language used throughout conflates the cell-type specificity conferred by the regulatory elements with that conferred by the serotype of the virus.

      Authors’ reply. In the revised version of the manuscript, we have corrected ambiguous language throughout.

      Recommendations for the authors

      Reviewer #1 (Recommendations For The Authors):

      My Public Review comments can be addressed by dialing down the interpretation of the data or providing appropriate caveats in the presentation of the relevant results and their discussion.

      We have done so. See text additions on p. 6 of the Results and p.10 of the Discussion.

      Minor comments:

      92% of PV+ neurons in the marmoset cortex were GABAergic. Can the authors speculate on the identity of the 8% PV+/GABA- neurons (e.g., on the basis of morphology)? Are they likely excitatory? Are they more likely to represent failures of GABA staining?

      We do not know what the other 8% of PV+/GABA- neurons are because we did not perform any other kind of IHC staining. Our best guess is that at least to some extent these represent failures of GABA staining, which is always challenging to perform in primate cortex. However, in mouse PV expression has been demonstrated in a minority of excitatory neurons.

      "Coverage of the PV-AAV was high, did not depend on injection volume.." The fact that the coverage did not depend on injection volume presumably depends, at least in part, on how ROIs were selected. Surely different volumes of injection transduce different numbers of neurons at different distances from the injection track. This should be clarified.

      The ROIs were selected at the center of the injected site/expression core from sections in which the expression region encompassed all cortical layers. Of course, larger volumes of injection resulted in larger transduced regions and therefore overall larger number of transduced neurons, but we counted cells only withing 100 µm wide ROIs at the center of the injection and the percent of transduced PV cells in this core region did not vary significantly across volumes. We have clarified the methods of ROI selection (see Methods pp. 13).

      Figure 2. What is meant by “absolute” in the legend for Figure 2? (How does “mean absolute density” differ from “mean density?”)

      We meant not relative, but this is obvious from the units, so we have removed the word “absolute” in the legend.

      Some non-significant p-values are indicated by "p>0.05" whereas others are given precisely (e.g., p = 1). Please provide precise p-values throughout. Also, the p-value from a surprisingly large number of comparisons in the first section of the results is "1". Is this due to rounding? Is it possible to get significance in a Bonferroni-corrected Kruskal-Wallis test with only 6 observations per condition?

      We now report exact p values throughout the manuscript (with a couple of exceptions where, in order to avoid reporting a large number of p values which interrupts the flow of the manuscript) we provide the upper bound value and state all those comparisons were below that value). The minimum sample size for Kruskall Wallis is 5, for each group being compared, and we our sample is 6 per group.

      Figure 3: The density of tdTomato-expressing cells appears to be greater at the AAV9 injection site than at the AAV1 injection site in the example sections shown. Might some of the differences between serotypes be due to this difference? I would imagine that resolving individual cells with certainty becomes more difficult as the amount of tdTomato expression increases.

      There was an error in the scale bar of Fig. 3C, so that the AAV1 injection site was shown at higher magnification than indicated by the wrong scale bar. Hence the density of tdTomato appeared lower than it is. Moreover, the tdT expression region shown in Fig. 3A is a merge of two sections, while it is only from a single section in panels B and C, leading to the impression of higher density of infected cells in panel A. The pipette used for the injection in panel A was not inserted perfectly vertical to the cortical surface, resulting in an injection site that did not span all layers in a single section; thus, to demonstrate that the injection indeed encompassed all layers (and that the virus infected cells in all layers), we collapsed label from two sections. We have now corrected the magnification of panel C so that it matches the scale bar in panel A, and specify in the figure legend that panel A label is from two sections.

      Text regarding Figure 3: The term “injection sizes” is confusing. I think it is intended to mean “the area over which tdTomato-expressing cells were found” but this should be clarified.

      Throughout the manuscript, we have changed the term injection site to “viral-expression region”.

      Figure 3: What were the titers of the three AAV-h56D vectors?

      Titers are now reported in the new Supplementary Table 1.

      Figure 3: The yellow box in Figure 3C is slightly larger than the yellow boxes in 3A and 3B. Is this an error or should the inset of Figure 3 have a scale bar that differs from the 50 µm scale bar in 3A?

      There were indeed errors in scale bars in this figure, which we have now corrected. Now all boxes have the same scale bar.

      Was MM423 one of the animals that received the AAV-h56D injections or one of the three that received AAV-S5E2 injection?

      This is an animal that received a 315nl injection of AAV-PHP.eB-S5E2.tdTomato. This is now specified in the Methods (see p. 12) and in the new Supplementary Table 1.

      Please provide raw cell counts and post-injection survival times for each animal.

      We now provide this information in Supplementary Tables 1 and 2.

      How were the different injection volumes of the AAV-S5E2 virus arranged by animal? Which volume of the AAV-S5E2 virus was injected into the two animals who received single injections?

      We now provide this information in Supplementary Table 1.

      Figure 6A: the point is made in the text that "[the distribution of tdT+ and PV+ neurons] did not differ significantly... peaking in L2/3 and 4C " Is the fact that the number of tdT+ and PV+ peak in layers 2/3 and 4C a consequence of these layers being thicker than the others? If so, this statement seems trivial.

      No, and this is the reason why we measured density in addition to percent of cells across layers in Figure 2. Figure 2B shows that even when measuring density, therefore normalizing by area, GABA+ and PV+ cell density still peaks in L2/3 and 4. Thus, these peaks do not simply reflect the greater thickness of these layers.

      Do the authors have permission to use data from Xu et al. 2010?

      Yes, we do.

      Reviewer #2 (Recommendations For The Authors):

      Minor comments:

      "Viral strategies to restrict gene expression to PV neurons have also been recently developed (Mehta et al., 2019; Vormstein-Schneider et al., 2020)." Mich et al. should also be cited here. Cell Rep. 2021;34(13):108754.

      We thank the reviewer for pointing out this missing references. This is now cited.

      “GABA density in L4C did not differ from any other layers, but the percent of GABA+ cells in L4C was significantly higher than in L1 (p=0.009) and 4A/B (p=<0.0001).” This and other similar observations depend on calculating the percentage of cells relative to the total number of DAPI-labeled cells in each layer. Since it is apparent that there must be considerable variability between layers, it would be helpful to add a histogram showing the densities of all DAPI-labeled cells for each layer.

      This is not how we calculated density. Density, as now clarified in the Results on p. 4, was defined as the number of cells per unit area. Counts in each layer were divided by each layers’ counting area. This corrects for differences in number of total labeled cells per layer. Therefore, reporting DAPI density is not necessary (we did not count DAPI cell density per layer).

      "Identical injection volumes of each serotype, delivered at 3 different cortical depths (see Methods), resulted in different injection sizes, suggesting the different serotypes have different capacity of infecting cortical neurons. AAV7 produced the smallest injection site, which additionally was biased to the superficial and deep layers, with only few cells expressing tdT in the middle layers (Fig. 3B). AAV9 (Fig. 3A) and AAV1 (Fig. 3C) resulted in larger injection sites and infected all cortical layers." Differences noted here might reflect either differences related to the AAV serotype or to differences in titers. Please add details about titers for each vector and add comments as appropriate. Another interpretation would be that there are differences in viral spread within the tissue.

      We have now added Supplementary Table 1 which reports titers in addition to other information about injections. The titers and volumes used for AAV9 and AAV7 were identical, while the titer for AAV1 was higher. Therefore, the differences in infectivity, particularly the much smaller expression region obtained with AAV7 cannot be attributed to titer. Likely this is due to differences in tropism and/or viral spread among serotypes. This is now discussed (see Results p. 5bottom and 6 top).

      “Recently, several viral vectors have been identified that selectively and efficiently restrict gene expression to GABAergic neurons and their subtypes across several species, but a thorough validation and characterization of these vectors in primate cortex has lacked.” Is this really a fair statement, or is the characterization presented here also lacking? Methods used by others for quantifying specificity and efficiency are essentially the same as used here. See for example Mich et al. (which is not cited).

      The original validation in primates of the vectors examined in our study was based on small tissue samples and did not examine the laminar expression profile of transgene expression induced by these enhancer-AAVs. For example, the validation of the h56D-AAV in marmoset cortex in the original paper by Mehta et al (2019) was performed on a tissue biopsy with no knowledge of which cortical layers were included in the tissue sample. The only study that shows laminar expression in primate cortex (Mich et al., which is now cited), only shows qualitative images of viral expression across layers, reporting total specificity and coverage pooled across samples; moreover, the study by Mich et al.  deals with different PV-specific enhancers than the ones characterized in our study. Unlike any of the previous studies, here we have quantified specificity and coverage across layers.

      "Specifically, we have shown that the GABA-specific AAV9-h56D (Mehta et al., 2019) induces transgene expression in GABAergic cells with up to 91-94% specificity and 80% coverage, and the PV-specific AAV-PHP.eB-S5E2 (Vormstein-Schneider et al., 2020) induces transgene expression in PV cells with up to 98% specificity and 86-90% coverage." These statements in the discussion repeat the somewhat exaggerated coverage numbers noted above for the Abstract.

      The averages across all layers are reported in the Results. The Discussion, abstract and discussion report upper limits, and this is made clear by stating “up to”, and now we have also added “depending on layer”.

      Reviewer #3 (Recommendations For The Authors):

      Abstract:

      • Ln 2: Can you be more specific about what you mean by the 'various functions of inhibition'? e.g. do you mean 'the various inhibitory influences on the local microcircuit' or similar?

      These are listed in the introduction to the paper but there is no space in the abstract to do so. Now the sentence reads: “various computational functions of…”.

      • Ln 5: 'has' to 'is'/'has been'.

      The grammar here is correct “has derived”.

      • Ln 6: humans are primates! Maybe change this to 'nonhuman primates'?

      We have added “non-human”

      • Ln n-1: 'viral vectors represent' -> 'viral vectors are'.

      We have changed it to “are”

      Intro:

      • Many readers may expect 'VIP' to be listed as the third major sub-class of interneurons. Could you note that the 5HT3a receptor-expressing group includes VIP cells?

      Done (p.3).

      • "Understanding cortical inhibitory neuron function in the primate is critical for understanding cortical function and dysfunction in the model system closest to humans" - this seems close to being circular logic (not quite, but close). Could you modify this sentence to reflect why understanding cortical function and dysfunction in NHP may be of interest?

      This sentence now reads (p.3):” Understanding cortical inhibitory neuron function in the primate is critical for understanding cortical function and dysfunction in the model system closest to humans, where cortical inhibitory neuron dysfunction has been implicated in many neurological and psychiatric disorders, such as epilepsy, schizophrenia and Alzheimer’s disease (Cheah et al., 2012; Verret et al., 2012; Mukherjee et al., 2019)”. We also note that this was already stated in the previous version of the paper but in the Discussion section which read (and still reads on p. 9 2nd paragraph): “It is important to study inhibitory neuron function in the primate, because it is unclear whether findings in mice apply to higher species, and inhibitory neuron dysfunction in humans has been implicated in several neurological and psychiatric disorders (Marin, 2012; Goldberg and Coulter, 2013; Lewis, 2014).”.

      • "In particular, two recent studies have developed recombinant adeno-associated viral vectors (AAV) that restrict gene expression to GABAergic neurons". This sentence places the emphasis on the wrong component of the technology. The fact that AAV was used is irrelevant; these constructs could equally have been packaged in a lenti, CAV, HSV, rabies, etc. The emphasis should be on the recently developed regulatory elements (the enhancers/promoters).

      Same problem with the following excerpts; this text implies that the serotype/vector confers cell-type selectivity, but the results presented do not support this assertion (the promoter/enhancer is what confers the selectivity).

      • "specifically, three serotypes of an AAV that restricts gene expression to GABAergic neurons".

      • "one serotype of an AAV that restricts gene expression to PV cells".

      • "GABA- and PV-specific AAVs".

      • "GABA-specific AAV" (in results).

      • "PV-specific AAVs".

      • "In this study, we have characterized several AAV vectors designed to restrict expression to GABAergic cells" (in discussion).

      • "GABA-virus". GABA is a NT, not a virus.

      We have modified the language in all these sections and throughout the manuscript.

      Results:

      • Enhancer specificity and efficiency cannot be directly compared, because they were packaged in different serotypes of AAV.

      We agree, and in fact we are not making comparisons between different enhancers (i.e., S5E2 and h56D).

      The three different serotypes of AAV expressing reporter under the h56D promoter were only tested once each, and all in the same animal. There are many variables that can contribute to the success (or failure) of a viral injection, so observations with an n=1 cannot be considered reliable.

      The authors need to either: (1) replicate the h56D virus injections in (at least) a second animal, or (2) rewrite the paper to focus on the AAV.PhP mDlx virus alone - for which they have adequate data - and mention the h56D data as an anecdotal result, with clear warnings about the preliminary nature of the observations due to lack of replication.

      We agree about the lack of sufficient data to make strong statements about the differences between serotypes for the h56D-AAV. In the revised version of the manuscript, following the Reviewers’ suggestion, we have chosen to temper our claims about differences between serotypes for the h56D enhancer and use more caution in the interpretation of these data. We feel that these data still demonstrate sufficiently high efficiency and specificity across cortical layers of transgene expression in GABA cells using the h56D promoter, at least with two of the 3 AAV serotypes we tested, to warrant their use in primates. Due to cost, time and ethical considerations related to the use of primates, we chose not to perform additional experiments to determine precise differences among serotypes. Thus, for example, while it is possible that had we replicated these experiments, serotype 7 could have proven equally efficient and specific as the other two serotypes, we felt answering this question did not warrant additional experiments in this precious species. Our edits in regard to this point can be found in the Results on p. 6 and Discussion on p. 10.

      • Did the authors compare h56D vs mDlx? This would be a useful and interesting comparison.

      We did not.

      • 3 tissue sections were used for analysis. How were these selected? Did the authors use a stereological approach?

      For the analysis in Fig. 2, the 3 sections were randomly selected and for the positioning of the ROIs we selected a region in dorsal V1 anterior to the posterior pole  (to avoid laminar distortions due to the curvature of the brain). This is now specified (see p. 4).

      • "both GABA+ and PV+ cells peak in layers" revise for clarity (e.g., the counts peak).

      In now reads “GABA+ and PV+ cell percent and density” (see p.4).

      • "we refer to this virus as GABA-AAV" these are 3 different viruses!

      The idea here was to use an abbreviation instead of using the full viral name every single time. Clearly the reviewer does not like this, so we have removed this convention throughout the paper and now specify the entire viral name each time.

      • "Identical injection volumes of each serotype, delivered at 3 different cortical depths (see Methods), resulted in different injection sizes". Do you mean 'resulted in different volumes of expression'?

      Yes. We have now rephrased this as follows: “…resulted in viral expression regions that differed in both size as well as laminar distribution” (p.5).

      • “suggesting the different serotypes have different capacity of infecting cortical neurons”. You can’t draw any firm conclusions from a single injection. The rest of this section of the results, along with the whole of Figure 4, and Figure 7a-d, is in danger of being misleading. Please remove. The best you can do here is to say ‘we injected 3 different viruses that express reporter under the h56D promoter. The results are shown in Figure 3, but these are anecdotal, as only a single injection of each virus was performed’. You could then note in the discussion to what extent these results are consistent with the existing literature (e.g., AAV9 often produces good coverage in NHP – anterograde and retrograde, AAV1 also works well in the CNS, although generally doesn’t infect as aggressively as AAV9. I’m not familiar with any attempts to use AAV7).

      With respect to Fig. 4, our approach in the revised version is detailed above. For convenience we copy it below here. With respect to Fig 7A-D, we feel the results are more robust as the data from the 3 serotypes here were pooled together, as the 3 serotype similarly downregulated GABA and PV expression at the injection site, and we do not make any statement about differences among serotypes for the data shown in Fig. 7A-D.

      “In the revised version of the manuscript, following the Reviewer ’s suggestion, we have chosen to temper our claims about differences between serotypes for the h56D enhancer and use more caution in the interpretation of these data (see revised text in the Results on p. 6 and in the Discussion on p. 10). We feel that these data still demonstrate sufficiently high efficiency and specificity across cortical layers of transgene expression in GABA cells using the h56D promoter, at least with two of the 3 AAV serotypes we tested, to warrant their use in primates. Due to cost, time and ethical considerations related to the use of primates, we chose not to perform additional experiments to determine precise differences among serotypes. Thus, for example, while it is possible that had we replicated these experiments, serotype 7 could have proven equally efficient and specific as the other two serotypes, we felt answering this question did not warrant additional experiments in this precious species.”

      • Figure 3: why the large variation in tissue quality? Are the 3 upper images taken at the same magnification? If not, they need different scale bars. The cells in A (upper row) look much smaller than those in B and C, and the size of the 'inset' box varies.

      We thank the reviewer for noticing this. We discovered an error in the scale bar of Fig. 3C, so that the AAV1 injection site was shown at higher magnification than indicated by the wrong scale bar. We have now corrected the error in scale bars. We have also fixed the different box sizes.

      • "Overall, across all layers coverage ranged from 78%{plus minus}1.9 for injection volumes >300nl to 81.6%{plus minus}1.8 for injection volumes of 100nl." Coverage didn't differ between layers, so revise this to: "Overall, across all layers coverage ranged from 78% to 81.6%." or give an overall mean (~80%).

      We have corrected the sentence as suggested by the Reviewer (see p. 8 first paragraph).

      • "extending farther from the borders" -> "extending beyond the borders".

      We have corrected the sentence as suggested by the Reviewer (see p. 8).

      • "The reduced GABA and PV immunoreactivity caused by the viruses implies that the specificity of the viruses we have validated in this study is likely higher than estimated". Yes, but for balance you should also note that they may harm the physiology of the cell.

      We have added a sentence acknowledging this to the Discussion. Specifically, on p. 10, we now state: “However, this reduced immunoreactivity raises concerns about the virus or high levels of reporter protein possibly harming the cell physiology.”

      Discussion:

      • "but a thorough validation and characterization of these vectors in primate cortex has lacked" better to say "has been limited", because Dimidschstein 2016 (marmoset V1) and Vormstein-schneider 2020 (macaque S1 and PFC) both reported expression in NHP.

      We have added the following sentence to this paragraph of the Discussion. “In particular, previous studies have not characterized the specificity and coverage of these vectors across cortical layers.”(see p. 8).

      • "whether finding in mice" -> 'whether findings in mice'.

      Corrected, thanks.

      • The discussion re: species differences is missing reference to Kreinen 2020 (10.1038/s41586-020-2781-z).

      This reference has been added. Thanks.

      • “Injections of about 200nl volume resulted in higher specificity (95% across layers) and coverage” – this is misleading. The coverage was not statistically different among injection volumes.

      We have added the following sentence: ”although coverage did not differ significantly across volumes.” (see p. 10).

      • "it is possible that subtle alteration of the cortical circuit upon parenchymal injection of viruses (including AAVs) leads to alteration of activity-dependent expression of PV and GABA." Or (and I would argue, more likely) the expression of large quantities of your big reporter protein compromised the function of the cell, leading to reduced expression of native proteins. You don't mention any IHC to amplify the RFP signal, so I'm assuming that your images are of direct expression. If so, you are expressing A LOT of reporter protein.

      We have added a sentence acknowledging this to the Discussion. Specifically, on p. 10, we now state: “However, this reduced immunoreactivity raises concerns about the virus or high levels of reporter protein possibly harming the cell physiology.”

      Methods:

      • It's difficult to piece together which viruses were injected in which monkeys, at what volumes, and at what titer. Please compile this info into a table for ease of reference (including any other relevant parameters).

      We now provide a Supplementary Table 1.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      The authors of this manuscript characterize new anion conducting that is more red-shifted in its spectrum than prior variants called MsACR1. An additional mutant variant of MsACR1 that is renamed raACR has a 20 nm red-shifted spectral response with faster kinetics. Due to the spectral shift of these variants, the authors proposed that it is possible to inhibit the expression of MsACR1 and raACR with lights at 635 nm in vivo and in vitro. The authors were able to demonstrate some inhibition in vitro and in vivo with 635 nm light. Overall the new variants with unique properties should be able to suppress neuronal activities with red-shifted light stimulation.

      Strengths:

      The authors were able to identify a new class of anion conducting channelrhodopsin and have variants that respond strongly to lights with wavelength >550 nm. The authors were able to demonstrate this variant, MsACR1, can alter behavior in vivo with 635 nm light. The second major strength of the study is the development of a red-shifted mutant of MsACR1 that has faster kinetics and 20 nm red-shifted from a single mutation.

      Weaknesses:

      The red-shifted raACR appears to work much less efficiently than MsACR1 even with 635 nm light illumination both in vivo (Figure 4) and in vitro (Figure 3E) despite the 20 nm red-shift. This is inconsistent with the benefits and effects of red-shifting the spectrum in raACR. This usually would suggest raACR either has a lower conductance than MsACR1 or that the membrane/overall expression of raACR is much weaker than MsACR1. Neither of these is measured in the current manuscript.

      Thank you for addressing this crucial issue. We posit that the diminished efficiency of raACR in comparison to MsACR1 WT can be attributed to the tenfold acceleration of its photocycle. As noted by Reviewer 1, the anticipated advantages associated with a red-shifted opsin, particularly in in vivo preparations, are offset by its accelerated off-kinetics. Consequently, the shorter dwell time of the open state leads to a reduced number of conducted ions per photon. Nevertheless, the operational light sensitivity is not drastically altered compared to MsACR WT (Fig. 3C). We believe that the rapid kinetics offer interesting applications, such as the precise inhibition of single action potentials through holography.

      There are limited comparisons to existing variants of ACRs under the same conditions in the manuscript overall. There should be more parallel comparison with gtACR1, ZipACR, and RubyACR in identical conditions in cultured cell lines, cultured neurons, and in vivo. This should be in terms of overall performance, efficiency, and expression in identical conditions. Without this information, it is unclear whether the effects at 635 nm are due to the expression level which can compensate for the spectral shift.

      We compared MsACR1 and raACR with GtACR1 in ND cells in supplemental figure 4. We concur that further comparisons could be useful to emphasise both the strengths of MsACRs and applications where they may not be as suitable. We are currently in the process of outlining a separate article. We firmly believe that each ACR variant occupies a distinct application niche, which necessitates a more comprehensive electrophysiological comparison to provide valuable insights to the scientific community.

      There should be more raw traces from the recordings of the different variants in response to short pulse stimulation and long pulse stimulation to different wavelengths. It is difficult to judge what the response would be like when these types of information are missing.

      We appreciate Reviewer 1's feedback and have compiled a collection of raw photoresponses, encompassing various pulse widths and wavelengths, which can be found in the Supplementary materials (Supplementary Figures 4 and 5).

      Despite being able to activate the channelrhodopsin with 635 nm light, the main utility of the variant should be transcranial stimulation which was not demonstrated here.

      We concur with Reviewer 1's assessment that MsACR prime application is indeed transcranial stimulation. However, it's worth emphasising that the full advantages of transcranial optical stimulation become most apparent when animals are truly freely moving without any tethered patch cords. Our ongoing research in the laboratory is dedicated to the development of a wireless LED system that can be securely affixed to the animal's skull. We aim to demonstrate the potential of these novell optogenetic approaches in the field of behavioural neuroscience in the coming year.

      Figure 3B is not clearly annotated and is difficult to match the explanation in the figure legend to the figure. The action potential spikings of neurons expressing raACR in this panel are inhibited as strongly as MsACR1.

      We have enhanced the figure caption and annotations for clarity. The traces presented in Figure 3B are intended to demonstrate the overall effectiveness of each variant. However, it is in the population data analysis, as depicted in Figure 3E, where the meaningful insights are revealed.

      For many characterizations, the number of 'n's are quite low (3-7).

      We acknowledge Reviewer 1's suggestion regarding the in vivo data and agree with the importance of including more animals, as well as control animals. However, we are committed to adhering to the principles of the 3Rs (Replacement, Reduction, Refinement) in animal research, and given the robustness of our observed effects, we will add animals to reach the minimal number of animals per condition (n = 2) to minimise unnecessary animal usage while ensuring statistical power.

      We will continue to adhere to the established standards in the field, aiming for a range of 3 to 7 cells per condition, sourced from at least two independent preparations, to ensure the robustness and reliability of our in vitro data.

      Reviewer #2 (Public Review):

      Summary:

      The authors identified a new chloride-conducting Channelrhodopsin (MsACR1) that can be activated at low light intensities and within the red part of the visible spectrum. Additional engineering of MsACR1 yielded a variant (raACR1) with increased current amplitudes, accelerated kinetics, and a 20nm red-shifted peak excitation wavelength. Stimulation of MsACR1 and raACR1 expressing neurons with 635nm in mice's primary motor cortices inhibited the animals' locomotion.

      Strengths:

      The in vitro characterization of the newly identified ACRs is very detailed and confirms the biophysical properties as described by the authors. Notably, the ACRs are very light sensitive and allow for efficient in vitro inhibition of neurons in the nano Watt/mm^2 range. These new ACRs give neuroscientists and cell biologists a new tool to control chloride flux over biological membranes with high temporal and spatial precision. The red-shifted excitation peaks of these ACRs could allow for multiplexed application with blue-light excited optogenetic tools such as cation-conducting channelrhodopsins or green-fluorescent calcium indicators such as GCaMP.

      Weaknesses:

      The in-vivo characterization of MsACR1 and raACR1 lacks critical control experiments and is, therefore, too preliminary. The experimental conditions differ fundamentally between in vitro and in vivo characterizations. For example, chloride gradients differ within neurons which can weaken inhibition or even cause excitation at synapses, as pointed out by the authors. Notably, the patch pipettes for the in vitro characterization contained low chloride concentrations that might not reflect possible conditions found in the in vivo preparations, i.e., increasing chloride gradients from dendrites to synapses.

      We appreciate Reviewer 2’s feedback regarding missing control experiments. We will respond to these concerns in another section of our manuscript, as suggested.

      Regarding the chloride gradient, we understand the concerns of Reviewer 2, yet we chose these ionic conditions, particularly as they were used in the initial electrical characterization of GtACR1 in a neuronal context (Mahn et al., 2016). We will make sure to provide this context in our manuscript to justify our choice of ionic conditions.

      Interestingly, the authors used soma-targeted (st) MsACR1 and raACR1 for some of their in vitro characterization yielding more efficient inhibition and reduction of co-incidental "on-set" spiking. Still, the authors do not seem to have utilized st-variants in vivo.

      At the time of submission, due to the long-term absence of our lab technician, we were not able to produce purified viruses. Therefore, we decided to move on with the submission. We now produced the virus externally, and will provide the experiments.

      Most importantly, critical in vivo control experiments, such as negative controls like GFP or positive controls like NpHR, are missing. These controls would exclude potential behavioral effects due to experimental artifacts. Moreover, in vivo electrophysiology could have confirmed whether targeted neurons were inhibited under optogenetic stimulations.

      We have several non-injected control animals that we used to calibrate this particular paradigm and never saw similar responses. However, we acknowledge the suggestion of Reviewer 2 and will include the GFP-injected control as recommended.

      Some of these concerns stem from the fact that the pulsed raACR stimulation at 635 nm at 10Hz (Fig. 3E) was far less efficient compared to MsACR1, yet the in vivo comparison yielded very similar results (Fig. 4D).

      As outlined previously, the accelerated photocycle of raACR results in a reduction in photocurrent amplitude, consequently diminishing the potency of inhibition per photon. In the context of in vitro stimulation, where single action potentials are recorded, this reduction in inhibition efficiency is resolved. However, in the realm of in vivo behavioural analysis, the observed effect is not contingent on single action potentials but rather stems from the disruption of the entire M1 motor network. In this context, despite the reduced efficiency of the fast-cycling raACR, it still manages to interrupt the M1 network, leading to similar behavioural outcomes.

      Also, the cortex is highly heterogeneous and comprises excitatory and inhibitory neurons. Using the synapsin promoter, the viral expression paradigm could target both types and cause differential effects, which has not been investigated further, for example, by immunohistochemistry. An alternative expression system, for example, under VGLUT1 control, could have mitigated some of these concerns.

      Indeed, we acknowledge the limitations of our current experimental approach. We are in the process of planning and conducting additional experiments involving cre-dependent expression of st-MSACR and st-raACR in PV-Cre mice.

      Furthermore, the authors applied different light intensities, wavelengths, and stimulation frequencies during the in vitro characterization, causing varying spike inhibition efficiencies. The in vivo characterization is notably lacking this type of control. Thus, it is unclear why the 635nm, 2s at 20Hz every 5s stimulation protocol, which has no equivalent in the in vitro characterization, was chosen.

      We appreciate the valuable comment from the reviewer. The objective of our in vitro characterization is to elucidate the general effects of specific stimulation parameters on the efficiency of neuronal inhibition. For instance, we aim to demonstrate that lower light intensities result in less efficient inhibition, or that pulse stimulation may lead to a less complete inhibition, albeit significantly reducing the energy input into the system.

      In the in vivo characterization, we face constraints such as animal welfare considerations and limitations in available laser lines, which prevent us from exploring the entire parameter space as comprehensively as in the in vitro preparation. Additionally, it is important to note that membrane capacitance tends to be higher in vivo compared to dissociated hippocampal neurons. Consequently, we have opted for a doubled stimulation frequency from 10 Hz to 20 Hz and the stimulation pattern of 2 seconds ”on” and 5 seconds “off”. This approach allows the animals to spend less time in an arrested state while still demonstrating the effect of MsACR and variants.

      In summary, the in vivo experiments did not confirm whether the observed inhibition of mouse locomotion occurred due to the inhibition of neurons or experimental artifacts.

      In addition, the author's main claim of more efficient neuronal inhibition would require them to threshold MsACR1 and raACR1 against alternative methods such as the red-shifted NpHR variant Jaws or other ACRs to give readers meaningful guidance when choosing an inhibitory tool.

      The light sensitivity of MsACR1 and raACR1 are impressive and well characterized in vitro. However, the authors only reported the overall light output at the fiber tip for the in vivo experiments: 0.5 mW. Without context, it is difficult to evaluate this value. Calculating the light power density at certain distances from the light fiber or thresholding against alternative tools such as NpHR, Jaws, or other ACRs would allow for a more meaningful evaluation.

      We thank the reviewers for their comments.

      Reviewer #1 (Recommendations For The Authors):

      The study would be much strengthened if the authors can perform more experiments and characterization to support their claims, in addition to showing more raw electrophysiological traces/results and not just summary charts and graphs.

      As outlined above, further experiments are planned. We appreciate the suggestion to include more raw electrophysiological traces. Photocurrent traces of all included mutants of MsACR1 measured in ND cells and traces of hippocampal neuronal measurements of non- and soma-targeted MsACR1 and raACR will be included as supplemental figures.

      Reviewer #2 (Recommendations For The Authors):

      Major concern:

      It is unclear if the optogenetic light stimulation in Fig. 4 caused direct inhibition of neuronal activity in M1, which cell types were targeted, and how MsACR1 and raACR1 compare to other optogenetic inhibitors.

      Also, the rationale for the light stimulation (635 nm, 2s, 20Hz, every 5s) is not clear.

      I would suggest the following to address these concerns:

      (1) M1 expression and stimulation of a negative control such as GFP to exclude that experimental artifacts cause the observed behavioral outcomes.

      We are now preparing the required GFP control, and will add it to the new version of the manuscript.

      (2) Expression and stimulation of NpHR as a positive control.

      We will use st-GtACR1 as a positive control.

      (3) Electrophysiological measurements of neuronal activity under optogenetic stimulation to confirm the effectiveness of neuronal inhibition, i.e. suppression of spontaneous firing under light etc.

      We concur with Reviewer 2 regarding the potential value of incorporating such in vivo optrode recordings into our manuscript to enable readers to assess the effectiveness of MsACR. As part of our plan for the next version of the manuscript, we intend to conduct these experiments.

      (4) ChR2 or other cation-conducting channelrhodopsins with the same expression paradigm could be used to observe diametrically opposite effects.

      As Reviewer 2 has already pointed out, the complex interactions that can occur in our viral strategy when an inhibitory opsin is expressed in both excitatory and inhibitory neurons make us sceptical about the possibility of an excitatory opsin leading to opposing effects.

      Considering the non-linear input-output function of cortical circuits, optogenetic activation of neurons, even when expressed in either inhibitory or excitatory neurons, is likely to result in the perturbation of the cortical network, which will likely also lead to locomotor arrest.

      (5) The authors should confirm whether the expression under synapsin preferentially targeted excitatory and inhibitory cells because inhibiting inhibitory cells could lead to the disinhibition of the principal cells. Synapsin promoters can drive expression in glutamatergic and GABAergic neurons. An alternative expression system under VGLUT1 promoter could yield better targeting.

      We concur with Reviewer 2 and will conduct the next set of experiments using the PV-Cre mouse line. Additionally, we will employ in vivo electrophysiology to further confirm the inhibition of the motor cortex network.

      (6) Titrating of optogenetic stimulation: The author should test whether increasing or decreasing light intensities and stimulation frequencies as well as different wavelengths (550 nm vs 635 nm) cause differences in inhibiting locomotion in vivo as it did for inhibiting the neuronal firing in vitro (Fig. 3B-E).

      The non-linear input-output function within cortical networks, coupled with our sole reliance on behaviour as a readout, will pose challenges in resolving subtle effects on locomotion arrest across various stimulation parameters.

      For our planned in vivo electrophysiology recordings, we will measure cortical firing rates as a proxy rather than relying solely on behavioural observations. This approach will allow us to map the fundamental axes of our parameter space in vivo, considering factors such as wavelength, light intensity, and frequency

      (7) Explanation of why the 20Hz/2s light stimulation protocol was chosen.

      As outlined above, considering animal welfare and increased membrane capacitance in vivo, we opted for the outlined stimulation protocol. This approach allows the animals to spend less time in an arrested state while still demonstrating the effect of MsACR and variants.

      (8) In vivo thresholding against other inhibitory tools, such as RubyACRs, Jaws, etc would provide critical guidance for the audience and potential users. It would be particularly important to compare the necessary light intensities for reaching similar behavioral outcomes.

      We concur with Reviewer 2 and will prepare data using GtACR1 as a reference.

      (9) The author should calculate or reasonably estimate the in vivo light intensity during optogenetic stimulation to provide a meaningful comparison to their in vitro characterization. Ideally, they can provide an estimated volume for efficient stimulation of MsACR1 and raACR1 and compare it to other optogenetic tools.

      We will conduct a Monte Carlo simulation and offer a comparison of the effective activation volume across various classes of optogenetic tools.

      Minor concerns:

      (1) Why were st- MsACR1 and raACR1 used in vitro but not in vivo? The viral constructs were described as AAV/DJ-hSyn1-MsACR-mCerulean and AAV/DJ-hSyn1-raACR-mCerulean.

      As mentioned earlier, we were unable to produce purified soma-targeted MsACR variants before the manuscript submission. We will now provide these measurements.

      (2) Light intensities for the spectral measurements are missing.

      During action spectra measurements, a motorised neutral density filter wheel is used to have equal photon flux for all tested wavelengths. Additionally, the light intensity is further reduced by using additional neutral density filters to ensure sufficiently low photocurrents to determine the spectral maximum. Therefore, the light intensity varied between constructs and sometimes measurements. We added the following line to the respective methods section to further clarify this: “(typically in the low µW-range at 𝜆max)”.

      (3) MsACR1 is slower and probably more light-sensitive than raACR1, which is faster but has larger photocurrents. These are complementary tradeoffs, and the audience might wonder how MsACR1 and raACR1 photocurrents compare under similar conditions. Therefore, I suggest an alternative representation in Fig. 2C. That is, the presentation of the excitation spectra under similar light intensities and with absolute photocurrent values.

      Unfortunately, due to the reasons stated above, MsACR1 and raACR action spectra were not recorded with the same light intensity. However, MsACR1 and raACR are compared under the same conditions for Fig. 2B, E, and F (560 nm light at ~3.2 mW/mm2) as well as in Supp. Fig. 4C.

      (4) Figure legends for figures 3F and G are missing details for describing the stimulation paradigm.

      We added more details about the stimulation paradigm.

    1. Reviewer #3 (Public Review):

      Summary:<br /> In this paper, the authors demonstrate the inevitably of the emergence of some degree of spatial information in sufficiently complex systems, even those that are only trained on object recognition (i.e. not "spatial" systems). As such, they present an important null hypothesis that should be taken into consideration for experimental design and data analysis of spatial tuning and its relevance for behavior.

      Strengths:<br /> The paper's strengths include the use of a large multi-layer network trained in a detailed visual environment. This illustrates an important message for the field: that spatial tuning can be a result of sensory processing. While this is a historically recognized and often-studied fact in experimental neuroscience, it is made more concrete with the use of a complex sensory network. Indeed, the manuscript is a cautionary tale for experimentalists and computational researchers alike against blindly applying and interpreting metrics without adequate controls.

      Weaknesses:<br /> However, the work has a number of significant weaknesses. Most notably: the degree and quality of spatial tuning is not analyzed to the standards of evidence historically used in studies of spatial tuning in the brain, and the authors do not critically engage with past work that studies the sensory influences of these cells; there are significant issues in the authors' interpretation of their results and its impact on neuroscientific research; the ability to linearly decode position from a large number of units is not a strong test of spatial information, nor is it a measure of spatial cognition; and the authors make strong but unjustified claims as to the implications of their results in opposition to, as opposed to contributing to, work being done in the field.

      The first weakness is that the degree and quality of spatial tuning that emerges in the network is not analyzed to the standards of evidence that have been used in studies of spatial tuning in the brain. Specifically, the authors identify place cells, head direction cells, and border cells in their network and their conjunctive combinations. However, these forms of tuning are the most easily confounded by visual responses, and it's unclear if their results will extend to forms of spatial tuning that are not. Further, in each case, previous experimental work to further elucidate the influence of sensory information on these cells has not been acknowledged or engaged with.

      For example, consider the head direction cells in Figure 3C. In addition to increased activity in some directions, these cells also have a high degree of spatial nonuniformity, suggesting they are responding to specific visual features of the environment. In contrast, the majority of HD cells in the brain are only very weakly spatially selective, if at all, once an animal's spatial occupancy is accounted for (Taube et al 1990, JNeurosci). While the preferred orientation of these cells are anchored to prominent visual cues, when they rotate with changing visual cues the entire head direction system rotates together (cells' relative orientation relationships are maintained, including those that encode directions facing AWAY from the moved cue), and thus these responses cannot be simply independent sensory-tuned cells responding to the sensory change) (Taube et al 1990 JNeurosci, Zugaro et al 2003 JNeurosci, Ajbi et al 2023).

      As another example, the joint selectivity of detected border cells with head direction in Figure 3D suggests that they are "view of a wall from a specific angle" cells. In contrast, experimental work on border cells in the brain has demonstrated that these are robust to changes in the sensory input from the wall (e.g. van Wijngaarden et al 2020), or that many of them are not directionally selective (Solstad et al 2008).

      The most convincing evidence of "spurious" spatial tuning would be the emergence of HD-independent place cells in the network, however, these cells are a small minority (in contrast to hippocampal data, Thompson and Best 1984 JNeurosci, Rich et al 2014 Science), the examples provided in Figure 3 are significantly more weakly tuned than those observed in the brain, and the metrics used by the authors to quantify place cell tuning are not clearly defined in the methods, but do not seem to be as stringent as those commonly used in real data. (e.g. spatial information, Skaggs et al 1992 NeurIPS).

      Indeed, the vast majority of tuned cells in the network are conjunctively selective for HD (Figure 3A). While this conjunctive tuning has been reported, many units in the hippocampus/entorhinal system are *not* strongly hd selective (Muller et al 1994 JNeurosci, Sangoli et al 2006 Science, Carpenter et al 2023 bioRxiv). Further, many studies have been done to test and understand the nature of sensory influence (e.g. Acharya et al 2016 Cell), and they tend to have a complex relationship with a variety of sensory cues, which cannot readily be explained by straightforward sensory processing (rev: Poucet et al 2000 Rev Neurosci, Plitt and Giocomo 2021 Nat Neuro). E.g. while some place cells are sometimes reported to be directionally selective, this directional selectivity is dependent on behavioral context (Markus et al 1995, JNeurosci), and emerges over time with familiarity to the environment (Navratiloua et al 2012 Front. Neural Circuits). Thus, the question is not whether spatially tuned cells are influenced by sensory information, but whether feed-forward sensory processing alone is sufficient to account for their observed turning properties and responses to sensory manipulations.

      These issues indicate a more significant underlying issue of scientific methodology relating to the interpretation of their result and its impact on neuroscientific research. Specifically, in order to make strong claims about experimental data, it is not enough to show that a control (i.e. a null hypothesis) exists, one needs to demonstrate that experimental observations are quantitatively no better than that control.

      Where the authors state that "In summary, complex networks that are not spatial systems, coupled with environmental input, appear sufficient to decode spatial information." what they have really shown is that it is possible to decode *some degree* of spatial information. This is a null hypothesis (that observations of spatial tuning do not reflect a "spatial system"), and the comparison must be made to experimental data to test if the so-called "spatial" networks in the brain have more cells with more reliable spatial info than a complex-visual control.

      Further, the authors state that "Consistent with our view, we found no clear relationship between cell type distribution and spatial information in each layer. This raises the possibility that "spatial cells" do not play a pivotal role in spatial tasks as is broadly assumed." Indeed, this would raise such a possibility, if 1) the observations of their network were indeed quantitatively similar to the brain, and 2) the presence of these cells in the brain were the only evidence for their role in spatial tasks. However, 1) the authors have not shown this result in neural data, they've only noticed it in a network and mentioned the POSSIBILITY of a similar thing in the brain, and 2) the "assumption" of the role of spatially tuned cells in spatial tasks is not just from the observation of a few spatially tuned cells. But from many other experiments including causal manipulations (e.g. Robinson et al 2020 Cell, DeLauilleon et al 2015 Nat Neuro), which the authors conveniently ignore. Thus, I do not find their argument, as strongly stated as it is, to be well-supported.

      An additional weakness is that linear decoding of position is not a strong test, nor is it a measure of spatial cognition. The ability to decode position from a large number of weakly tuned cells is not surprising. However, based on this ability to decode, the authors claim that "'spatial' cells do not play a privileged role in spatial cognition". To justify this claim, the authors would need to use the network to perform e.g. spatial navigation tasks, then investigate the network's ability to perform these tasks when tuned cells were lesioned.

      Finally, I find a major weakness of the paper to be the framing of the results in opposition to, as opposed to contributing to, the study of spatially tuned cells. For example, the authors state that "If a perception system devoid of a spatial component demonstrates classically spatially-tuned unit representations, such as place, head-direction, and border cells, can "spatial cells" truly be regarded as 'spatial'?" Setting aside the issue of whether the perception system in question does indeed demonstrate spatially-tuned unit representations comparable to those in the brain, I ask "Why not?" This seems to be a semantic game of reading more into a name then is necessarily there. The names (place cells, grid cells, border cells, etc) describe an observation (that cells are observed to fire in certain areas of an animal's environment). They need not be a mechanistic claim (that space "causes" these cells to fire) or even, necessarily, a normative one (these cells are "for" spatial computation). This is evidenced by the fact that even within e.g. the place cell community, there is debate about these cells' mechanisms and function (eg memory, navigation, etc), or if they can even be said to serve only a single function. However, they are still referred to as place cells, not as a statement of their function but as a history-dependent label that refers to their observed correlates with experimental variables. Thus, the observation that spatially tuned cells are "inevitable derivatives of any complex system" is itself an interesting finding which *contributes to*, rather than contradicts, the study of these cells. It seems that the authors have a specific definition in mind when they say that a cell is "truly" "spatial" or that a biological or artificial neural network is a "spatial system", but this definition is not stated, and it is not clear that the terminology used in the field presupposes their definition.

      In sum, the authors have demonstrated the existence of a control/null hypothesis for observations of spatially-tuned cells. However, 1) It is not enough to show that a control (null hypothesis) exists, one needs to test if experimental observations are no better than control, in order to make strong claims about experimental data, 2) the authors do not acknowledge the work that has been done in many cases specifically to control for this null hypothesis in experimental work or to test the sensory influences on these cells, and 3) the authors do not rigorously test the degree or source of spatial tuning of their units.

    1. Reviewer #1 (Public Review):

      This manuscript by Kleinman & Foster investigates the dependence of hippocampal replay on VTA activity. They recorded neural activity from the dorsal CA1 region of the hippocampus while chemogenetically silencing VTA dopamine neurons as rats completed laps on a linear track with reward delivery at each end. Reward amount changed across task epochs within a session on one end of the track. The authors report that VTA activity is necessary for an increase in sharp-wave rate to remain localized to the feeder that undergoes a change in reward magnitude, an effect that was especially pronounced in a novel environment. They follow up on this result with a second experiment in which reward magnitude varies unpredictably at one end of the linear track and report that changes in sharp-wave rate at the variable location reflect both the amount of reward rats just received there, in addition to a smaller modulation that is reminiscent of reward prediction error coding, in which the previous reward rats received at the variable location affects the magnitude of the subsequent change in sharp-wave rate that occurs on the present visit.

      This work is technically innovative, combining neural recordings with chemogenetic inactivation. The question of how VTA activity affects replay in the hippocampus is interesting and important given that much of the work implicating hippocampal replay in memory consolidation and planning comes from reward-motivated behavioral tasks. Enthusiasm for the manuscript is dampened by some technical considerations about the chemogenetic portion of the experiments. Additionally, there are some interpretational issues related to whether changes in reward magnitude affected sharp-wave rate directly, or whether the reported changes in sharp-wave rate alter behavior and these behavioral changes affect sharp-wave rate.

      Major issues:

      Chemogenetics validation

      Little validation is provided for the chemogenetic manipulations. The authors report that animals were excluded due to lack of expression but do not quantify/document the extent of expression in the animals that were included in the study. There's no independent verification that VTA was actually inhibited by the chemogenetic manipulation besides the experimental effects of interest.

      The authors report a range of CNO doses. What determined the dose that each rat received? Was it constant for an individual rat? If not, how was the dose determined? The authors may wish to examine whether any of their CNO effects were dependent on dose.

      The authors tested the same animal multiple times per day with relatively little time between recording sessions. Can they be certain that the effect of CNO wore off between sessions? Might successive CNO injections in the same day have impacted neural activity in the VTA differently? Could the chemogenetic manipulation have grown stronger with each successive injection (or maybe weaker due to something like receptor desensitization)? The authors could test statistically whether the effects of CNO that they report do not depend on the number of CNO injections a rat received over a short period of time.

      Motivational considerations

      In a similar vein, running multiple sessions per day raises the possibility that rats' motivation was not constant across all data collection time points. The authors could test whether any measures of motivation (laps completed, running speed) changed across the sessions conducted within the same day. This is a particularly tricky issue, because my read of the methods is that saline sessions were only conducted as the first session of any recording day, which means there's a session order/time of day and potential motivational confound in comparing saline to CNO sessions.

      Statistics, statistical power, and effect sizes

      Throughout the manuscript, the authors employ a mixture of t-tests, ANOVAs, and mixed-effects models. Only the mixed effects models appropriately account for the fact that all of this data involves repeated measurements from the same subject. The t-tests are frequently doubly inappropriate because they both treat repeated measures as independent and are not corrected for multiple comparisons.

      The number of animals in these studies is on the lower end for this sort of work, raising questions about whether all of these results are statistically reliable and likely to generalize. This is particularly pronounced in the reward volatility experiment, where the number of rats in the experimental group is halved to just two. The results of this experiment are potentially very exciting, but the sample size makes this feel more like pilot data than a finished product.

      The effect sizes of the various manipulations appear to be relatively modest, and I wonder if the authors could help readers by contextualizing the magnitude of these results further. For instance, when VTA inactivation increases mis-localization of SWRs to the unchanged end of the track, roughly how many misplaced sharp-waves are occurring within a session, and what would their consequence be? On this particular behavioral task, it's not clear that the animals are doing worse in any way despite the mislocalization of sharp-waves. And it seems like the absolute number of extra sharp-waves that occur in some of these conditions would be quite small over the course of a session, so it would be helpful if the authors could speculate on how these differences might translate to meaningful changes in processes like consolidation, for instance.

      How directly is reward affecting sharp-wave rate?

      Changes in reward magnitude on the authors' task cause rats to reallocate how much time they spent at each end. Coincident with this behavioral change, the authors identify changes in the sharp-wave rate, and the assumption is that changing reward is altering the sharp-wave rate. But it also seems possible that by inducing longer pauses, increased reward magnitude is affecting the hippocampal network state and creating an occasion for more sharp-waves to occur. It's possible that any manipulation so altering rats' behavior would similarly affect the sharp-wave rate.

      For instance, in the volatility experiment, on trials when no reward is given sharp-wave rate looks like it is effectively zero. But this rate is somewhat hard to interpret. If rats hardly stopped moving on trials when no reward was given, and the hippocampus remained in a strong theta network state for the full duration of the rat's visit to the feeder, the lack of sharp-waves might not reflect something about reward processing so much as the fact that the rat's hippocampus didn't have the occasion to emit a sharp-wave. A better way to compute the sharp-wave rate might be to use not the entire visit duration in the denominator, but rather the total amount of time the hippocampus spends in a non-theta state during each visit. Another approach might be to include visit duration as a covariate with reward magnitude in some of the analyses. Increasing reward magnitude seems to increase visit duration, but these probably aren't perfectly correlated, so the authors might gain some leverage by showing that on the rare long visit to a low-reward end sharp-wave rate remains reliably low. This would help exclude the explanation that sharp-wave rate follows increases in reward magnitude simply because longer pauses allow a greater opportunity for the hippocampus to settle into a non-theta state.

      The authors seem to acknowledge this issue to some extent, as a few analyses have the moments just after the rat's arrival at a feeder and just before departure trimmed out of consideration. But that assumes these sorts of non-theta states are only occurring at the very beginning and very end of visits when in fact rats might be doing all sorts of other things during visits that could affect the hippocampus network state and the propensity to observe sharp-waves.

      Minor issues

      The title/abstract should reflect that only male animals were used in this study.

      The title refers to hippocampal replay, but for much of the paper the authors are measuring sharp-wave rate and not replay directly, so I would favor a more nuanced title.

      Relatedly, the interpretation of the mislocalization of sharp-waves following VTA inactivation suggests that the hippocampus is perhaps representing information inappropriately/incorrectly for consolidation, as the increased rate is observed both for a location that has undergone a change in reward and one that has not. However, the authors are measuring replay rate, not replay content. It's entirely possible that the "mislocalized" replays at the unchanged end are, in fact, replaying information about the changed end of the track. A bit more nuance in the discussion of this effect would be helpful.

      The authors use decoding accuracy during movement to determine which sessions should be included for decoding of replay direction. Details on cross-validation are omitted and would be appreciated. Also, the authors assume that sessions failed to meet inclusion criteria because of ensemble size, but this information is not reported anywhere directly. More info on the ensemble size of included/excluded sessions would be helpful.

      For most of the paper, the authors detect sharp-waves using ripple power in the LFP, but for the analysis of replay direction, they use a different detection procedure based on the population firing rate of recorded neurons. Was there a reason for this switch? It's somewhat difficult to compare reported sharpwave/replay rates of the analyses given that different approaches were used.

    1. Nina Paley’s Sita Sings The Blues, released online a little over two months ago, has been generating great press and even greater viewership, closing in on 70,000 downloads at archive.org alone. For the non-inundated, there is great background information on the film at Paley’s website. We recently had the opportunity to talk with Paley about the film – we touched on the film’s aesthetics and plot points, but perhaps most interesting to those in the CC community is Paley’s decision to utilize our copyleft license, Attribution-ShareAlike, and her thoughts on free licensing and the open source movement in general. Read on to learn more about the licensing trials and tribulations associated with the film’s release, how CC has played a role, and Paley’s opinions on the Free Culture movement as a whole. RamSitaGods, Nina Paley | CC BY-SA One of the major stories surrounding Sita Sings The Blues been your use of songs by musician Annette Hanshaw and the back-and-forth dialogue you have had with the copyright owners as a result. Can you explain why you used these songs? The songs themselves inspired the film. There would be no film without those songs. Until I heard them, the Ramayana was just another ancient Indian epic to me. I was feebly connecting this ancient epic to my own experiences in 2002. But the Hanshaw songs were a revelation: Sita’s story has been told a million times not just in India, not just through the Ramayana, but also through American Blues. Hers is a story so primal, so basic to human experience, it has been told by people who never heard of the Ramayana. The Hanshaw songs deal with exactly the same themes as the epic; but they emerged completely independent of it. Their sound is distinctively 1920’s American, and therein lies their power: the listener/viewer knows I didn’t make them up. They are authentic. They are historical evidence supporting the film’s central point: the story of the Ramayana transcends time, place and culture. What is this story? Sita is a goddess/princess/woman utterly devoted to her husband Rama, the god/prince/man. Sita’s story moves from total enmeshment and romantic joy (Here We Are, What Wouldn’t I Do For That Man) to hopeful longing separation (Daddy Won’t You Please Come Home) to reunion (Who’s That Knockin’ At My Door) to romantic rejection (Mean to Me) to reconciliation (If You Want the Rainbow) to further rejection (Moanin’ Low, Am I Blue) to hopeless longing (Lover Come Back to Me,) back to love – this time self-love (I’ve Got a Feelin’ I’m Fallin’). Sita’s role is to suffer, especially through loving a man who rejects her. Women especially connect emotionally to her story and these emotions are clearly expressed in songs. As Nabaneeta Dev Sen writes in “Lady sings the Blues: When Women retell the Ramayana”: But there are always alternative ways of using a myth. If patriarchy has used the Sita myth to silence women, the village women have picked up the Sita myth to give themselves a voice. They have found a suitable mask in the myth of Sita, a persona through which they can express themselves, speak of their day-to-day problems, and critique patriarchy in their own fashion. Sen is talking about the songs of Indian village women, but she could just as easily been talking about American Blues. That is the point of Sita Sings the Blues: we all struggle with this story, which connects humans through time, space and culture, whether we’re aware of it or not. Just as the Ramayana has mostly been written down and controlled by men, the songs in Sita Sings the Blues were mostly written by men; but sung by a woman – Hanshaw – they pack an emotional wallop and express a woman’s voice. The synchronicity of the Hanshaw songs and Sita’s story is uncanny. This impresses audiences and allows the film’s point to be made: the story of the Ramayana transcends time, place and culture. Because the songs feature an authentic voice from the 1920’s, they demonstrate that this story emerged organically in history. New songs composed by the director, while they could be entertaining, could not make that point. They would be a mere contrivance, whereas the authentic, historical songs give weight to the film’s thesis. They are in fact the basis of the film’s thesis, irrefutable evidence that certain stories – like the story of Sita and Rama – are inherent to human experience. Upon reading the above, Karl Fogel added: Using something that already exists demonstrates that the universality of your theme is external to yourself. Whereas causing something new to exist wouldn’t achieve the same effect. Instead, it would be circular: it would demonstrate that the artist has the ability to make more of what she’s already making. So rather than being connective or expanding, it would be narcissistic (just in a descriptive sense, not necessarily a pejorative one). There has to be a reason so many composers, even non-Catholic ones like Bach, set the Latin Mass to music instead of making up their own words. (Hmm, now imagine if those words had been monopoly-restricted… 🙂 ). What has your experience been in trying to get permission it use Hanshaw’s music in the film, and the current state of affairs? Because distributors were going bankrupt right and left in 2008, it was no longer possible to sell an indie film to a distributor for big money and then “have them take care of” the licenses. Since in February of 2008, when the film premiered in Berlin, I was not yet a Free Culture convert, I thought I needed a conventional distributor. So it fell on me to clear the rights. I had to pay intermediaries to contact the license holders, since they don’t speak to mere riff raff like me; they’re too busy, and under no obligation to do so. Even before that, I needed legal help to research who owned the rights in the first place, since there’s no central copyright registry any more, and rights are traded like baseball cards between corporations. Luckily, I was aided by the student attorneys of the Glushko-Samuelson Intellectual Property Law Clinic of American University. Anyway, in 2008 a lawyer charged me $7,000 to get this response from the licensors: an estimate of $15,000 to $26,000 per song, AFTER I’d paid a $500 per song Festival License. (Festival Licenses last one whole year and require a promise to not make any money showing the film. So a festival license isn’t enough to get the “week-long commercial run” required for Academy Award qualification. Now that “Sita”‘s been broadcast, she will never qualify for an Academy nomination; if I’d really wanted one, I would have had to delayed the release of the film for another year. But I digress.). Even though we made it explicitly clear the entire budget for the film was under $200,000, the licensors came back with the “bargain” estimate of about $220,000. It was simply not possible for me to acquire that kind of money. So legally, my only option was to not show the film or commit civil disobedience. I hired another intermediary, a “rights clearance house” which is less expensive than a lawyer, and they negotiated the “step deal” I eventually signed. This brought the price tag of the licenses down to $50,000, but with many restrictions. If more than 5,000 DVDs (or downloads) are sold, I must pay the licensors more. I wrote about this at length on my website. I borrowed $50,000 to pay these licenses for several reasons. First, to reduce my liability. I may still be sued for releasing the film freely online – after all, the licensors may interpret free sharing as “selling” for zero dollars – but I’ll only be sued for breach of contract, not copyright infringement. Copyright infringement carries much harsher penalties, including possible jail time. I also wanted to make free sharing of “Sita” as legal, and therefore legitimate, as possible. Sharing shouldn’t be the exclusive purview of lawbreakers. Sharing should – and can – be wholesome fun for the whole family. I paid up to indemnify the audience, because the audience is Sita’s main distributor. So it’s now legal to copy and share Sita Sings the Blues. The files went up on Archive.org in early March 2009 and have spread far and wide since. Having paid off the licensors, I could have chosen conventional distribution. But I chose a CC BY-SA license to allow the film to reach a much wider audience; to prohibit the copyrighting – “locking up” – of my art; to give back to the greater culture which gave to me; to exploit the power of the audience to promote and distribute more efficiently than a conventional distributor; and to educate about the dangers of copy restrictions, and the beauty and benefits of sharing. As a result of the trouble you’ve had in regards to Annete Hanshaw’s music, you have turned into a self-proclaimed Free Culture activist. Was this shift gradual? What has that experience in particular informed your views on copyright, fair use, and the public domain? Annette Hanshaw was immensely popular in the late 1920’s. Now almost no one’s heard of her. Why? Because of copy-restrictions. I met many talented filmmakers on my “festival circuit.” Most had conventional distribution deals, but it’s very hard to see any of their films, which had small, brief theatrical runs, and then were never heard from again. Why? Copy-restrictions. I’m an artist. I need money to live, but even more importantly I need my art to reach people. A $10,000 advance in return for having my work locked up for 10 years is a devil’s bargain. More than anything, I wanted people to see my film – now and in years to come. My turning point in choosing a CC license happened in October of 2008. “Sita” had just opened the San Francisco Animation Festival, and I’d disclosed to the audience we’d all just done something illegal. It’s always great to share the film on a big screen in a theater with an audience, and this one was particularly enthusiastic. The next morning I woke up realizing that a free release online wouldn’t in any way prevent theatrical screenings. Why had I never considered that before? Because the film industry insists people won’t go to theaters if they can see a film online. But that’s not true of me, nor many cinephiles. When I lived in San Francisco my favorite movie outings were to classic films at the Catsro: 2001, Nights of Cabiria, Modern Times, Mommy Dearest. These are all available on home video, but I went to the Castro for the big screen and the dark room and the shared experience. If enough people watched and liked “Sita” online, there’d be demand for it in cinemas. And so far that’s proving true. In particular, how have you viewed CC licenses in this whole process? What was your motivation to release Sita Sings the Blues under a CC BY-SA license? Why did you choose that license and not another CC license? What are the obstacles and benefits you’ve seen in using CC licenses? I want my film to reach the widest audience. It costs money to run a theater; it costs money to manufacture DVDs; it costs money to make and distribute 35mm film prints. It’s essential I allow people to make money distributing Sita these ways and others; otherwise, no one will do it. So I eschewed the “non commercial” license. Share Alike would “protect” the work from ever being locked up. It’s better than Public Domain; works are routinely removed from the Public Domain via privatized derivatives (just try making your own Pinocchio). I didn’t want some corporation locking up a play or TV show based on Sita. They are certainly welcome to make derivative works, and make money from them; in fact I encourage this. But they may not sue or punish anyone for sharing those works. I looked to the Free Software movement as a model. The CC BY-SA license most closely resembles the GNU GPL, which is the foundation of Free Software. People make plenty of money in Free Software; there’s no reason they can’t do the same in Free Culture, except for those pernicious “non commercial” licenses. A Share Alike license eliminates the corporate abuse everyone’s so afraid of, while it encourages entrepreneurship and innovation. Everyone wins, especially the artist! What else would you like our reader’s to know? Any plans for the future? I’d love you all to read my essay Understanding Free Content and of course watch the film! I’m currently busy making “containers” like DVDs and T shirts available now at our e-store. QuestionCopyright is my main partner in releasing Sita; we’re trying to prove a model in which freedom and revenue work together. We know other filmmakers are watching what happens to Sita, and we’d like to show that yes, you can make money without impinging on everyone else’s freedom. I’m also negotiating with theatrical distributors in France and Switzerland, as well as a couple book publishers. I’m negotiating not “rights” to the film, which belong to everyone already, but rather my Endorsement and assistance. To understand how this works, please read about the Creator Endorsed Mark. Once I have the Sita Sings the Blues Merchandise Empire started, I hope to work on short musical cartoons about free speech – you can hear one of the songs here. There’s more where that came from. Really, I have more ideas than I have time to implement them – a happy yet vexing problem. I also hope to have all my old Nina’s Adventures and Fluff syndicated comic strips scanned and uploaded at high resolution onto archive.org under a CC BY-SA license. The University of Illinois Library is currently seeking funding to move ahead on this project – interested individuals should contact Betsy Kruger. Lastly, I’m still looking for money, although the Sita Sings the Blues Merchandise Empire should be generating some in a few months. Still, I plan to apply for grants and fellowships. Any foundations with too much money burning a hole in your accounts, please get in touch.

      In this text, it dives into how Ramayana as a text is so universal that any set of tunes or music can match it. For example, this text looks into how the music from Annette Hanshaw from the 1920's are able to blend into the Indian epic showing its versatility. One challenge that readers might resonate to is the copyright issues that Paley faced in order to have permission to use Hanshaw's music since there were many legal problems and a bunch of fees. Because of this struggle, it highlights why it can be difficult to use certain words in conjunction with other pieces which might explain why we might not see the types of works that we would like. Even looking at the copyright license that Paley chose for her own film, she chose the one that would allow her to reach a larger group of people because her goal is not to make money but to appreciate art for what it is and to share that with other people. The copyright restrictions that are discussed in this text can be eye-opening for a lot of readers as they can see why creativity might be hindered in a lot of fields and this can help explain why. In this text, the concepts of culture and national identity are closely related to the idea of self. This can be seen in Ramayana as its themes are universal and the ability for it to mesh well with the American Blues songs is proof of that. Not to mention, this is an example that serves to show that cultures can blend together in which a person's self can be a multitude of different aspects reflecting in how modern nation-building does not just rely on one perspective or facet, but it can have many different facets allowing that identity to be fluid as a result. The Indians relating to Ramayana may see themselves as "us" because they resonate to that Indian epic while "them" represents those who know more about the American Blues or western culture in general. With this being said, this contrast in cultures being able to blend and mesh well together show how there is shared human experiences across cultures. There is no sense of otherness in Paley's work because she is able to show how the themes in Ramayana are universal and can be applied to all time periods and all locations. Even though the argument is that Ramayana is universal and can blend with any music types, the choice of American Blues is compelling by Paley and this intention was because she may have wanted to see how English lyrics can mesh with Sanskrit language as a challenge and this weird combination can prove that it would work with all music types as a result. Because of this contrast, it speaks to the power of linguistic authenticity as it is able to prove the themes behind the film and put them into action. The difficulty that Paley faced with copyright laws help explain why people cannot be as creative as they want and why free sharing should exist. As a result, Paley allows her work to be easily more accessed which can be seen in her creative commons license and shows that she backs up the same claim from her own film as well. It shows why artists and all people should move away from exclusive ownership and should embrace a more collaborative model in which all people can contribute and take inspirations from each other in positive ways. CC BY Ajey Sasimugunthan (contact)

    1. Author response:

      The following is the authors’ response to the original reviews.

      Reviewer #1 (Public Review): 

      Summary: This article explores the role of Ecdysone in regulating female sexual receptivity in Drosophila. The researchers found that PTTH, throughout its role as a positive regulator of ecdysone production, negatively affects the receptivity of adult virgin females. Indeed, loss of larval PTTH before metamorphosis significantly increases female receptivity right after adult eclosion and also later. However, during metamorphic neurodevelopment, Ecdysone, primarily through its receptor EcR-A, is required to properly develop the P1 neurons since its silencing led to morphological changes associated with a reduction in adult female receptivity. Nonetheless, the result shown in this manuscript sheds light on how Ecdysone plays a dual role in female adult receptivity, inhibiting it during larval development and enhancing it during metamorphic development. Unfortunately, this dual and opposite effect in two temporally different developmental stages has not been highlighted or explained. 

      Strengths: This paper exhibits multiple strengths in its approach, employing a well-structured experimental methodology that combines genetic manipulations, behavioral assays, and molecular analysis to explore the impact of Ecdysone on regulating virgin female receptivity in Drosophila. The study provides clear and substantial findings, highlighting that removing PTTH, a positive Ecdysone regulator, increases virgin female receptivity. Additionally, the research expands into the temporal necessity of PTTH and Ecdysone function during development. 

      Weaknesses: 

      There are two important caveats with the data that are reflecting a weakness: 

      (1) Contradictory Effects of Ecdysone and PTTH: One notable weakness in the data is the contrasting effects observed between Ecdysone and its positive regulator PTTH. PTTH loss of function increases female receptivity, while ecdysone loss of function reduces it. Given that PTTH positively regulates Ecdysone, one would expect that the loss of function of both would result in a similar phenotype or at least a consistent directional change. 

      A1. As newly formed prepupae, the ptth-Gal4>UAS-Grim flies display similar changes in gene expression to the genetic control flies to response to a high-titer ecdysone pulse. These include the repression of EcR (McBrayer et al.,2007). We tested whether there is a similar feedforward relationship between PTTH and EcR-A. We quantified the EcR-A mRNA level of PTTH -/- and PTTH -/+ in the whole body of newly formed prepupae. Indeed, PTTH -/- induced increased EcR-A expression in the whole body of newly formed prepupae compared with PTTH -/+ flies. Because of the function of EcR-A in gene expression, this suggests that PTTH -/- disturbs the regulation of a serious of gene expressions during metamorphosis. However, it is not sure that the EcR-A expression in pC1 neurons is increased compared with genetic controls when PTTH is deleted. Furthermore, PTTH -/- must affect development of other neurons rather than only pC1 neurons. So, the feedforward relationship between PTTH and EcRA at the start of prepupal stage is one possible cause for the contradictory effects of PTTH -/- and EcR-A RNAi in pC1 neurons.  

      (2) Discordant Temporal Requirements for Ecdysone and PTTH: Another weakness lies in the different temporal requirements for Ecdysone and PTTH. The data from the manuscript suggest that PTTH is necessary during the larval stage, as shown in Figure 2 E-G, while Ecdysone is required during the pupal stage, as indicated in Figure 5 I-K. Ecdysone is a crucial developmental hormone with precisely regulated expression throughout development, exhibiting several peaks during both larval and pupal stages. PTTH is known to regulate Ecdysone during the larval stage, specifically by stimulating the kinetics of Ecdysone peaking at the wandering stage. However, it remains unclear whether pupal PTTH, expressed at higher levels during metamorphosis, can stimulate Ecdysone production during the pupal stage. Additionally, given the transient nature of the Ecdysone peak produced at wandering time, which disappears shortly before the end of the prepupal stage, it is challenging to infer that larval PTTH will regulate Ecdysone production during the pupal stage based on the current state of knowledge in the neuroendocrine field.  

      Considering these two caveats, the results suggest that the authors are witnessing distinct temporal and directional effects of Ecdysone on virgin female receptivity.  

      A2. First of all, it is necessary to clarify the detailed time for the manipulation of Ptth gene and PTTH neurons. In Figure 3, activation of PTTH neurons during the stage 2 inhibited the female receptivity. The “stage 2” is from six hours before the 3rd-instar larvae to the end of the wandering larvae (the start of prepupae). In Figure 5, The “pupal stage” is from the prepupal stage to the end of pupal stage. This “pupal stage” includes the forming of prepupae when the ecdysone peak is not disappeared. The time of manipulating Ptth and EcR-A in pC1 neurons are continuous. In addition, the pC1-Gal4 expressing neurons appear also at the start of prepupal stage. So, it is possible that PTTH regulates female receptivity through the function of EcR-A in pC1 neurons. 

      Reviewer #1 (Recommendations For The Authors): 

      In light of the significant caveat previously discussed, I will just make a few general suggestions: 

      (1) The paper primarily focuses on robust phenotypes, particularly in PTTH mutants, with a well-detailed execution of several experiments, resulting in thorough and robust outcomes. However, due to the caveat previously presented (opposite effect in larva and pupa), consider splitting the paper into two parts: Figures 1 to 4 deal with the negative effect of PTTH-Ecdysone on early virgin female receptivity, while Figures 5 to 7 focus on the positive metamorphic effect of Ecdysone in P1 metamorphic neurodevelopment. However, in this scenario, the mechanism by which PTTH loss of function increases female receptivity should be addressed.

      A3. It is a good suggestion that splitting the paper into two parts associated with the PTTH function and EcR function in pC1 neurons separately, if it is impossible that PTTH functions in female receptivity through the function of EcR-A in pC1 neurons. However, because of the feedforward relationship between PTTH and EcR-A in the newly formed prepupae, and the time of manipulating Ptth and EcR-A in pC1 neurons is continuous, it is possible that these two functions are not independent of each other. So, we still keep the initial edition.

      (2) Validate the PTTH mutants by examining homozygous mutant phenotypes and the dose-dependent heterozygous mutant phenotype using existing PTTH mutants. This could also be achieved using RNAi techniques.

      A4. We did not get other existing PTTH mutants. We instead decreased the PTTH expression in PTTH neurons and dsx+ neurons, but did not detect the similar phenotype to that of PTTH -/-. Similarly, the overexpression through PTTH-Gal4>UAS-PTTH is also not sufficient to change female receptivity. It is possible that both decreasing and increasing PTTH expression are not sufficient to change female receptivity.

      (3) Clarify if elav-Gal4 is not expressed in PTTH neurons and discuss how the rescue mechanisms work (hormonal, paracrine, etc.) in the text.

      A5. We tested the overlap of elav-Gal4>GFP signal and the stained PTTH with PTTH antibody. We did not detect the overlap. It suggests that elav-Gal4 is not expressed in PTTH neurons. However, we detected the expression of PTTH (PTTH antibody) in CNS when overexpressed PTTH using elav-Gal4>UASPTTH based on PTTH -/-. Furthermore, this rescued the phenotype of PTTH -/- in female receptivity. Insect PTTH isoforms have similar probable signal peptide for secreting. Indeed, except for the projection of axons to PG gland, PTTH also carries endocrine function acting on its receptor Torso in light sensors to regulate light avoidance of larvae. The overexpressed PTTH in other neurons through elav-Gal4>UASPTTH may act on the PG gland through endocrine function and then induce the ecdysone synthesis and release. So that, although elav-Gal4 is not expressed in PTTH neurons, the ecdysone synthesis triggered by PTTH from the hemolymph may result in the rescued PTTH -/- phenotype in female receptivity.

      (4) Consider renaming the new PTTH mutant to avoid confusion with the existing PTTHDelta allele. 

      A6. We have renamed our new PTTH mutant as PtthDelete.

      (5) Include the age of virgin females in each figure legend, especially for Figures 2 to 7, to aid in interpretation. This is essential information since wild-type early virgins -day 1- show no receptivity. In contrast, they reach a typical 80% receptivity later, and the mechanism regulating the first face might differ from the one occurring later.

      A7. We have included the age of virgin females in each figure legend. 

      (6) Explain the relevance of observing that PTTH adult neurons are dsx-positive, as it's unclear why this observation is significant, considering that these neurons are not responsible for the observed receptivity effect in virgin females. Alternatively, address this in the context of the third instar larva or clarify its relevance.  

      A8. We decreased the DsxF expression in PTTH neurons and did not detect significantly changed female receptivity. Almost all neurons regulating female receptivity, including pC1 neurons, express DsxF. We suppose that PTTH neurons have some relationship with other DsxF-positive neurons which regulate female receptivity. Indeed, we detected the overlap of dsx-LexA>LexAop-RFP and torso-Gal4>UAS-GFP during larval stage. Furthermore, decreasing Torso expression in pC1 neurons significantly inhibit female receptivity. 

      These results suggest that, PTTH regulates female receptivity not only through ecdysone, but also may through regulating other neurons especially DsxF-positive neurons associated with female receptivity directly. 

      Reviewer #2 (Public Review): 

      Summary: The authors tried to identify novel adult functions of the classical Drosophila juvenile-adult transition axis (i.e. ptth-ecdysone). Surprisingly, larval ptth-expressing neurons expressed the sex-specific doublesex gene, thus belonging to the sexual dimorphic circuit. Lack of ptth during late larval development caused enhanced female sexual receptivity, an effect rescued by supplying ecdysone in the food. Among many other cellular players, pC1 neurons control receptivity by encoding the mating status of females. Interestingly, during metamorphosis, a subtype of pC1 neurons required Ecdysone Receptor A in order to regulate such female receptivity. A transcriptomic analysis using pC1-specific Ecdyone signaling down-regulation gives some hints of possible downstream mechanisms. 

      Strengths: the manuscript showed solid genetic evidence that lack of ptth during development caused enhanced copulation rate in female flies, which includes ptth mutant rescue experiments by overexpressing ptth as well as by adding ecdysone-supplemented food. They also present elegant data dissecting the temporal requirements of ptth-expressing neurons by shifting animals from non-permissive to permissive temperatures, in order to inactivate neuronal function (although not exclusively ptth function). By combining different drivers together with a EcR-A RNAi line authors also identified the Ecdysone receptor requirements of a particular subtype of pC1 neurons during metamorphosis. Convincing live calcium imaging showed no apparent effect of EcR-A in neural activity, although some effect on morphology is uncovered. Finally, bulk RNAseq shows differential gene expression after EcR-A down-regulation. 

      Weaknesses: the paper has three main weaknesses. The first one refers to temporal requirements of ptth and ecdysone signaling. Whereas ptth is necessary during larval development, the ecdysone effect appears during pupal development. ptth induces ecdysone synthesis during larval development but there is no published evidence about a similar role for ptth during pupal stages. Furthermore, larval and pupal ecdysone functions are different (triggering metamorphosis vs tissue remodeling). The second caveat is the fact that ptth and ecdysone loss-of-function experiments render opposite effects (enhancing and decreasing copulation rates, respectively). The most plausible explanation is that both functions are independent of each other, also suggested by differential temporal requirements. Finally, in order to identify the effect in the transcriptional response of down-regulating EcR-A in a very small population of neurons, a scRNAseq study should have been performed instead of bulk RNAseq. 

      In summary, despite the authors providing convincing evidence that ptth and ecdysone signaling pathways are involved in female receptivity, the main claim that ptth regulates this process through ecdysone is not supported by results. More likely, they'd rather be independent processes. 

      B1. Clarification: in Figure 3, activation of PTTH neurons during the stage 2 inhibited the female receptivity. The “stage 2” is from six hours before the 3rd-instar larvae to the end of the wandering larvae (the start of prepupae). In Figure 5, The “pupal stage” is from the start of prepupal stage to the end of pupal stage. This “pupal stage” includes the forming of prepupae when the ecdysone peak is not disappeared. The time of manipulating Ptth and EcR-A in pC1 neurons are continuous. In addition, the pC1-Gal4 expressing neurons appear also at the start of prepupal stage. So, it is possible that PTTH regulates female receptivity through the function of EcR-A in pC1 neurons. 

      B2. During the forming of prepupae, the ptth-Gal4>UAS-Grim flies display similar changes in gene expression to the genetic control flies to response to a high-titer ecdysone pulse. These include the repression of EcR (McBrayer et al.,2007). We tested whether there is a similar feedforward relationship between PTTH and EcR-A. We quantified the EcR-A mRNA level of PTTH -/- and PTTH -/+ in the whole body of newly formed prepupae. Indeed, PTTH -/- induced increased EcR-A compared with PTTH -/+ flies. Because of the function of EcR-A in gene expression, this suggests that PTTH -/- disturbs the regulation of a serious of gene expressions during metamorphosis. However, it is not sure that the EcR-A expression in pC1 neurons is increased compared with genetic controls when PTTH is deleted. Furthermore, PTTH -/- must affect the development of other neurons rather than only pC1 neurons. So, the feedforward relationship between PTTH and EcR-A at the start of prepupal stage is one possible cause for the contradictory effects of PTTH -/- and EcR-A RNAi in pC1 neurons.

      B3. We will do single cell sequencing in pC1 neurons for the exploration of detailed molecular mechanism of female receptivity in the future.

      Reviewer #2 (Recommendations For The Authors): 

      Additional experiments and suggestions: 

      - torso LOF in the PG to determine whether or not the ecdysone peak regulated by ptth (there is a 1-day delay in pupation) is responsible for the ptth effect in L3. In the same line, what happens if torso is downregulated in the pC1 neurons? Is there any effect on copulation rates? 

      B4. Because the loss of phm-Gal4, we could not test female receptivity when decreasing the expression of Torso in PG gland. However, decreasing Torso expression in pC1 neurons significantly inhibit female receptivity. This suggests that PTTH regulates female receptivity not only through ecdysone but also through regulating dsx+ pC1 neurons in female receptivity directly.

      - What is the effect of down-regulating ptth in the dsx+ neurons? No ptth RNAi experiments are shown in the paper. 

      B5. We decreased PTTH expression in dsx+ neurons but did not detect the change in female receptivity.  We also decreased PTTH expression in PTTH neurons using PTTH-Gal4, also did not detect the change in female receptivity. Similarly, the overexpression through PTTH-Gal4>UAS-PTTH is also not sufficient to change female receptivity. It is possible that both decreasing and increasing PTTH expression are not sufficient to change female receptivity.

      - Why are most copulation rate experiments performed between 4-6 days after eclosion? ptth LOF effect only lasts until day 3 after eclosion (but very weak-fig 1). Again, this supports the idea that ptth and ecdysone effects are unrelated.

      B6. Most behavioral experiments were performed between 4-6 days after eclosion as most other studies in flies, because the female receptivity reaches the peak at that time. Ptth LOF made female receptivity enhanced from the first day after eclosion. This seems like the precocious puberty. Wild type females reach high receptivity at 2 days after eclosion (about 75% within 10 min). We suppose that Ptth LOF effect only lasts until day 3 after eclosion because too high level of receptivity of control flies to exceed.

      It is not sure whether the effect of PTTH-/- in female receptivity disappears after the 3rd day of adult flies. So that it is not sure whether PTTH and EcR-A effects in pC1 neurons are unrelated.

      - The fact that pC1d neuronal morphology changes (and not pC1b) does not explain the effect of EcR-A LOF. Despite it is highlighted in the discussion, data do not support the hypothesis. How do these pC1 neurons look like in a ptth mutant animal regarding Calcium imaging and/or morphology? 

      B7. We detected the pattern of pC1 neurons when PTTH is deleted. Consistent with the feedforward relationship between PTTH and expression of EcR-A in newly formed prepupae, PTTH deletion induced less established pC1-d neurons contrary to that induced by EcR-A reduction in pC1 neurons. However, it is not sure that the expression of EcR-A in pC1 neurons is increased when PTTH is deleted. Furthermore, on the one hand, manipulation of PTTH has general effect on the neurodevelopment not only regulating pC1 neurons. On the other hand, the detailed pattern of pC1-b neurons which is the key subtype regulating female receptivity when EcR-A is decreased in pC1 neurons or PTTH is deleted could not be seen clearly. So, the abnormal development of pC1-b neurons, if this is true, is just one of the possible reasons for the effect of PTTH deletion on female receptivity.

      - The discussion is incomplete, especially the link between ptth and ecdysone; discuss why the phenotype is the opposite (ptth as a negative regulator of ecdysone in the pupa, for instance); the difference in size due to ptth LOF might be related to differential copulation rates.  

      B8. We have revised the discussion. We could not exclude the effect of size of body on female receptivity when PTTH was deleted or PTTH neurons were manipulated, although there was not enough evidence for the effect of body size on female receptivity.

      - scheme of pC neurons may help. 

      B9. We have tried to label pC1 neurons with GFP and sort pC1 neurons through flow cytometry sorting, but could not success. This may because the number of pC1 neurons is too low in one brain. We will try single-cell sequencing in the future. 

      - Immunofluorescence images are too small.

      B10. We have resized the small images.

      Reviewer #3 (Public Review): 

      Summary: 

      This manuscript shows that mutations that disable the gene encoding the PTTH gene cause an increase in female receptivity (they mate more quickly), a phenotype that can be reversed by feeding these mutants the molting hormone, 20-hydoxyecdysone (20E). The use of an inducible system reveals that inhibition or activation of PTTH neurons during the larval stages increases and decreases female receptivity, respectively, suggesting that PTTH is required during the larval stages to affect the receptivity of the (adult) female fly. Showing that these neurons express the sex-determining gene dsx leads the authors to show that interfering with 20E actions in pC1 neurons, which are dsx-positive neurons known to regulate female receptivity, reduces female receptivity and increases the arborization pattern of pC1 neurons. The work concludes by showing that targeted knockdown of EcRA in pC1 neurons causes 527 genes to be differentially expressed in the brains of female flies, of which 123 passed a false discovery rate cutoff of 0.01; interestingly, the gene showing the greatest down-regulation was the gene encoding dopamine beta-monooxygenase. 

      Strengths 

      This is an interesting piece of work, which may shed light on the basis for the observation noted previously that flies lacking PTTH neurons show reproductive defects ("... females show reduced fecundity"; McBrayer, 2007; DOI 10.1016/j.devcel.2007.11.003). 

      Weaknesses: 

      There are some results whose interpretation seem ambiguous and findings whose causal relationship is implied but not demonstrated. 

      (1) At some level, the findings reported here are not at all surprising. Since 20E regulates the profound changes that occur in the central nervous system (CNS) during metamorphosis, it is not surprising that PTTH would play a role in this process. Although animals lacking PTTH (rather paradoxically) live to adulthood, they do show greatly extended larval instars and a corresponding great delay in the 20E rise that signals the start of metamorphosis. For this reason, concluding that PTTH plays a SPECIFIC role in regulating female receptivity seems a little misleading, since the metamorphic remodeling of the entire CNS is likely altered in PTTH mutants. Since these mutants produce overall normal (albeit larger--due to their prolonged larval stages) adults, these alterations are likely to be subtle. Courtship has been reported as one defect expressed by animals lacking PTTH neurons, but this behavior may stand out because reduced fertility and increased male-male courtship (McBrayer, 2007) would be noticeable defects to researchers handling these flies. By contrast, detecting defects in other behaviors (e.g., optomotor responses, learning and memory, sleep, etc) would require closer examination. For this reason, I would ask the authors to temper their statement that PTTH is SPECIFICALLY involved in regulating female receptivity.  

      C1. We agree with that, it is not surprising that PTTH regulates the profound changes that occur in the CNS during metamorphosis through ecdysone. Also, the behavioral changes induced by PTTH mutants include not only female receptivity. We will temper the statement about the function of PTTH on female receptivity.

      We think there are two new points in our text although more evidences are needed in the future. On the one hand, PTTH deletion and the reduction of EcR-A in pC1 neurons during metamorphosis have opposite effects on female receptivity. On the other hand, development of pC1-b neurons regulated by EcR-A during metamorphosis is important for female receptivity.

      (2) The link between PTTH and the role of pC1 neurons in regulating female receptivity is not clear. Again, since 20E controls the metamorphic changes that occur in the CNS, it is not surprising that 20E would regulate the arborization of pC1 neurons. And since these neurons have been implicated in female receptivity, it would therefore be expected that altering 20E signaling in pC1 neurons would affect this phenotype. However, this does not mean that the defects in female receptivity expressed by PTTH mutants are due to defects in pC1 arborization. For this, the authors would at least have to show that PTTH mutants show the changes in pC1 arborization shown in Fig. 6. And even then the most that could be said is that the changes observed in these neurons "may contribute" to the observed behavioral changes. Indeed, the changes observed in female receptivity may be caused by PTTH/20E actions on different neurons.

      C2. As newly formed prepupae, the ptth-Gal4>UAS-Grim flies display similar changes in gene expression to the genetic control flies to response to a high-titer ecdysone pulse. These include the repression of EcR (McBrayer et al., 2007). We tested whether there is a similar feedforward relationship between PTTH and EcR-A. We quantified the EcR-A mRNA level of PTTH -/- and PTTH -/+ in the whole body of newly formed prepupae. Indeed, PTTH -/- induced upregulated EcR-A in the whole body of newly formed prepupae compared with PTTH -/+ flies. We also detected the pattern of pC1 neurons when PTTH is deleted. Consistent with the feedforward relationship between PTTH and expression of EcR-A in newly formed prepupae, PTTH deletion induced less established pC1-d neurons contrary to that induced by EcR-A reduction in pC1 neurons. 

      However, it is not sure that the expression of EcR-A in pC1 neurons increases compared with genetic controls when PTTH is deleted. Furthermore, on the one hand, manipulation of PTTH has general effect on the neurodevelopment. On the other hand, the detailed pattern of pC1-b neurons which is the key subtype regulating female receptivity through EcR-A function in pC1 neurons could not be seen clearly. So, the abnormal development of pC1b neurons, if this is true, is just one of the possible reasons for the effect of PTTH deletion on female receptivity.

      (3) Some of the results need commenting on, or refining, or revising:  a- For some assays PTTH behaves sometimes like a recessive gene and at other times like a semidominant, and yet at others like a dominant gene. For instance, in Fig. 1D-G, PTTH[-]/+ flies behave like wildtype (D), express an intermediate phenotype (E-F), or behave like the mutant (G). This may all be correct but merits some comment.

      C3. Female receptivity increases with the increase of age after eclosion, not only for wild type flies but also PTTH mutants. At the first day after eclosion (Figure 1D), maybe the loss of PTTH in PTTH[-]/+ flies is not enough for sexual precocity as in PTTH -/-. At the second day after eclosion and after (Figure 1E-G), the loss of PTTH in PTTH[-]/+ flies is sufficient to enhance female receptivity compared with wild type flies. However, After the 2nd day of adult, female receptivity of all genotype flies increases sharply. At the 3rd day of adult and after, female receptivity of PTTH -/- reaches the peak and the receptivity of PTTH[-]/+ reaches more nearly to PTTH -/- when flies get older.  

      b - Some of the conclusions are overstated. i) Although Fig. 2E-G does show that silencing the PTTH neurons during the larval stages affects copulation rate (E) the strength of the conclusion is tempered by the behavior of one of the controls (tub-Gal80[ts]/+, UAS-Kir2.1/+) in panels F and G, where it behaves essentially the same as the experimental group (and quite differently from the PTTH-Gal4/+ control; blue line).(Incidentally, the corresponding copulation latency should also be shown for these data.). ii) For Fig. 5I-K, the conclusion stated is that "Knock-down of EcR-A during pupal stage significantly decreased the copulation rate." Although strictly correct, the problem is that panel J is the only one for which the behavior of the control lacking the RNAi is not the same as that of the experimental group. Thus, it could just be that when the experiment was done at the pupal stage is the only situation when the controls were both different from the experimental. Again, the results shown in J are strictly speaking correct but the statement is too definitive given the behavior of one of the controls in panels I and K. Note also that panel F shows that the UAS-RNAi control causes a massive decrease in female fertility, yet no mention is made of this fact.

      C4. i) For all figures in the text, only when all the control groups were significant different from assay group, we say the assay group is significantly different. In Figure 2E-G, the control groups were both different from the assay group only at the larval stage. The difference between two control groups may due to the genetic background. We have described more detailed statistical analysis in the legend. In addition, the corresponding copulation latency has been shown. ii) For Figure 5, we have revised the conclusion in text as “when the experiment was done at the pupal stage is the only situation when the controls were both different from the experimental.” Besides, the UAS-RNAi control causes a massive decrease in female fertility in panel F has been mentioned.

      Reviewer #3 (Recommendations For The Authors): 

      (1) I am not sure that PTTH neurons should be referred to as "PG neurons". I am aware that this name has been used before but the PG is a gland that does not have neurons; it is not even innervated in all insects. 

      C5. Agree. “PG neurons” has been changed into “PTTH neurons”.

      (2) Fig. 1A warrants some explanation. One can easily imagine what it shows but a description is warranted. 

      C6. Explanation has been added.

      (3) When more than one genotype is compared it would be more useful to use letters to mark the genotypes that are not statistically different from each other rather than simply using asterisks. For instance, in the case of copulation latencies shown in Fig. 1E-G, which result does the comparison refer to? For example, since the comparisons are the result of ANOVAs, which comparison receives "*" in Fig. 1F? Is it PTTH[-]/+ vs PTTH[-]/PTTH[-] or vs. +/+? 

      C7. Referred genotypes and conditions were marked in all figure legends.

      (4) Fig. 1H: Why is copulation latency of PTTH[-]/PTTH[-]+elav-GAL4 significantly different from that of PTTH[-]/PTTH[-]? This merits a comment. Also, why was elav-GAL4 used to effect the rescue and not the PTTH-GAL4 driver? 

      C8. We could not explain this phenomenon. This may due to the different genetic backgrounds between controls. We have mentioned this in figure legend.

      (5) Fig. 2C, the genotype is written in a confusing order, GAL4+UAS should go together as should LexA+LexAop. 

      C9. We have revised for avoiding confusion.

      (6) In Fig. 2, is "larval stage" the same period that is shown in Fig. 3A? Please clarify.

      C10. We have clarified this in text and legends.

      (7) Fig. 6. The fact that pC1 neurons can be labeled using the pC1-ss2-Gal4 at the start of the pupal stage does not mean that this is when these neurons appear (are born), only when they start expressing this GAL4. Other types of evidence would be needed to make a statement about the birthdate of these neurons. 

      C11. We have revised the description for the appearance of pC1-ss2-Gal4>GFP. The detailed birth time of pC1 neurons will be tested in future.

      (8) The results shown in Fig. 7 are not pursued further and thus appear like a prelude to the next manuscript. Unless the authors have more to add regarding the role of one of the differentially expressed genes (e.g., dopamine beta-monooxygenase, which they single out) I would suggest leaving this result out. 

      C12. We have leave this out.

      (9) Female flies lacking PTTH neurons were reported to show lower fecundity by McBrayer et al. (2007) and should be cited. 

      C13. This important study has been cited in the first manuscript. In this revision, we have cited it again when mentioning the lower fecundity of female flies lacking PTTH neurons.

      (10) Line 230: when were PTTH neurons activated? Since they are dead by 10h post-eclosion it isn't clear if this experiment even makes sense. 

      C14. Yes, we did this for making sure that PTTH neurons do not affect female receptivity at adult stage again.

      (11) Line 338: the statements in the figures say that PTTH function is required during the larval stages, not during metamorphosis 

      C15. This has been revised as “The result suggested that EcR-A in pC1 neurons plays a role in virgin female receptivity during metamorphosis. This is consistent with that PTTH regulates virgin female receptivity before the start of metamorphosis.”

      (12) Did the authors notice any abnormal behavior in males? McBrayer et al. (2007) mention that males lacking PTTH neurons show male-male courtship. This may remit to the impact of 20E on other dsx[+] neurons. 

      C16. Yes, we have noticed that males lacking PTTH show male-male courtship. It is possible that PTTH deletion induces male-male courtship through the impact of 20E on other dsx+ or fru+ neurons. We have added the corresponding discussion.

      (13) Line 145: please define CCT at first use 

      C17. CCT has been defined.

      (14) Overall the manuscript is well written; however, it would still benefit from editing by a native English speaker. I have marked a few corrections that are needed, but I probably missed some. 

      + Line 77: "If female is not willing..." should say "If THE female is not willing..." 

      + Line 78 "...she may kick the legs, flick the wings," should say "...she may kick HER legs, flick HER wings," 

      + Lines 93-94 this sentence is unclear: "...while the neurons in that fru P1 promoter or dsx is expressed regulate some aspects..." 

      + Line 108 "...similar as the function of hypothalamic-pituitary-gonadal (HPG).." should say "...similar

      TO the function of hypothalamic-pituitary-gonadal (HPG).." 

      + Line 152 "Due to that 20E functions through its receptor EcR.." should say ""BECAUSE 20E ACTS through its receptor EcR.." 

      + Lines 155, 354 "unnormal" is not commonly used (although it is an English word); "abnormal" is usually used instead. 

      + Line 273: "....we then asked that whether ecdysone regulates" delete "that"  + Sentences lines 306-309 need to be revised.

      C18. Thank you for your suggestions. We have revised as you advise.

    1. your communications skills help you to understand others—not just their words, but also their tone of voice, their nonverbal gestures, or the format of their written documents provide you with clues about who they are and what their values and priorities may be.

      I would also add that another means of communication is through facial expressions. If someone is interested in the conversation or disgusted by it, it's obvious. Despite the fact that most facial expressions contradict ideal gestures, some people are able to distinguish between them.

    1. Note: This response was posted by the corresponding author to Review Commons. The content has not been altered except for formatting.

      Learn more at Review Commons


      Reply to the reviewers

      Reviewer #1 (Evidence, reproducibility and clarity (Required)): Summary: To explore the relationship between histone post-translational modifications (H3K4me3 and H3K27me3) and enhancer activation with gene expression during early embryonic development, the authors used a monolayer differentiation approach to convert mouse embryonic stem cells (ESCs) into Anterior Definitive Endoderm (ADE). They monitored differentiation stages using a dual reporter mESC line (B6), which has fluorescent reporters inserted at the Gsc (GFP) and Hhex (Redstar) loci. Their analyses indicate that the differentiating cells advanced through stages similar to those in the embryo, successfully converting into endoderm and ADE with high efficiency. This is elegant and well performed stem cell biology.

      Their subsequent genome-wide and nascent transcription analyses confirmed that the in vitro gene expression changes correlated with developmental stages and confirmed that transcriptional activation precedes mRNA accumulation. They then focussed on linking active enhancers and histone modifications (H3K4me3 and H3K27me3) were with gene expression dynamics. Finally, the performed PRC2 inhibition and showed that, while it enhanced differentiation efficiency, it also induced ectopic expression of non-lineage specific genes.

      Major comments: In terms of mechanistic advances, they propose that transcriptional up-regulation does not require prior loss of H3K27me3, which they show appears to lag behind gene activation, but critically, on a likely mixed population level. I am sceptical of their interpretation of their data because they are looking at heterogenous populations of cells. To explain, one could imagine a particular H3K27me3 coated gene that gets activated during differentiation. In a population of differentiating cells, while the major sub-population of cells could retain H3K27me3 on this particular gene when it is repressed, a minority sub-population of cells could have no H3K27me3 on the gene when it is actively transcribed. The ChIP and RNA-seq results in this mixed cell scenario would give the wrong impression that the gene is active while retaining H3K27me3, when in reality, it's much more likely that the gene is never expressed when its locus in enriched with the repressive H3K27me3 modification. Therefore, to support their claim, they would have to show that a particular gene is active when its locus is coated with H3K27me3. Personally, I don't feel this approach would be worth pursuing.

      They also report that inhibition of PRC2 using EZH2 inhibitor (EPZ6438) enhanced endoderm differentiation efficiency but led to ectopic expression of pluripotency and non-lineage genes. However, this is not surprising considering the established role of Polycomb proteins as repressors of lineage genes.

      Reviewer #1 (Significance (Required)): I feel that this is a solid and well conducted study in which the authors model early development in vitro. It should be of interest to researchers with an interest in more sophisticated in vitro differentiation systems, perhaps to knockout their gene of interest and study the consequences. However, I don't see any major mechanistic advances in this work.

      *>Author Response *

      *We agree with the point regarding the delayed loss of H3K27me3 relative to gene activation, and indeed this same point has been raised by reviewer 3 (see below). Our cell-population based data does not allow us to directly test if gene up-regulation in a small population of cells from TSSs lacking H3K27me3, accounts for the observed result. Furthermore, there are currently no robust methods to determine cell- or allele-specific expression simultaneously with ChIP/Cut and Run for chromatin marks. However, we provide the following additional evidence that strongly supports our conclusions. *

      • *

      Our FACs isolation strategy used to prepare cell populations for ChIP, microarray expression and 4sU-seq analysis is based on expression (or lack thereof) of a fluorescent GSC-GFP reporter. This means that every cell in the G+ populations express the Gsc fluorescent reporter, at least at the protein level, at the point of isolation. This is despite the presence of appreciable and invariant levels of H3K27me3 at the TSS of the Gsc gene in both G+ and G- populations at day 3 of differentiation. Comparable to our meta-analysis of all upregulated genes shown in the original manuscript (Figure 5 and S5), H3K27me3 levels are then subsequently reduced in the G+ relative to the G- populations at day 4. The transcriptional changes which correspond to the GSG-GFP reporter expression and associated ChIP-seq data are shown in the reviewer figure (Fig R1 A shown in revision plan). To further support our observations, we sought to rule out the possibility that the shift in H3K27me3 and transcription were from mutually exclusive gene sets, from nominal transcription levels or from sites with low level H3K27me3. To do this with a gene set of sufficient size to yield a robust result, we selected upregulated TSSs that had a greater than median value for both transcription (4sU-seq) and H3K27me3 (n=49 of 159 genes; Fig R1 B shown in revision plan). Meta-analysis of these genes showed that, as for all upregulated gene TSS (n=159), transcriptional activation occurred in the presence of substantial and invariant levels of H3K27me3 at day 3 followed by a subsequent reduction by day 4 of differentiation (Fig R1 C shown in revision plan). Importantly, many of these genes yielded high absolute 4sU-seq signal, comparable to that of Gsc, arguing against transcriptional activation being limited to a small subpopulation of cells.

      • *

      Reviewer #2 (Evidence, reproducibility and clarity (Required)): In this paper the authors profile gene expression, including active transcription, and histone modifications (k4 and k27me3) during a complex differentiation protocol from ES cells, which takes advantage of FACS sorting of appropriate fluorescent reporters. The data is of good quality and the experiments are well performed. The main conclusion, that the analyzed histone marks channel differentiation more than they directly allow/block it, is well supported by the data. The paper is interesting and will represent a good addition to an already extensive literature. I have however a few major concerns, described below:

      1/ K4me3 may show more changes than they interpret, at least over the +1 nucl. An alternative quantification to aggregate profiles should be used to more directly address the questions regarding the correlations between histone mods and gene expression.

      *>Author Response *

      *Whilst we state that H3K4me3 levels are somewhat invariant at differentially expressed genes relative to H3K27me3, quantification of individual TSS (+/- 500 bp) did show a direct correlation with gene expression (Figure 5 and S5). To further explore this in response to the reviewer’s comment we will quantify K4me3 signal at the +1 nucleosome to determine if this yields more substantial differences than that observed more broadly across TSSs. *

      2/ Related to the previous point, it appears clear in Fig.4 that the promoters of each gene expression cluster do not belong to a single chromatin configuration. I think it would be important to: 1/ cluster the genes based on promoter histone mods and interrogate gene expression and cluster allocation (basically the reverse to what is presented) 2/ order the genes in the heatmaps identically for K4me3 and K27me3 to more easily understand the respective chromatin composition per cluster

      >Author Response

      We thank the reviewer for these suggestions and will include these analyses in a revised manuscript.

      3/ Also, as it is apparent that not all promoters in every cluster are enriched for the studied marks, could the authors separately analyze these genes? What are they? Do they use alternative promoters?

      >Author Response

      *Indeed, this is the case. Whilst there is significant enrichment of H3K27me3 at the TSS of developmentally regulated genes, not all genes whose expression changes during the differentiation will be polycomb targets. We will further stratify these clusters as suggested and determine what distinguishes the subsets. If informative, this data will be included in a revised manuscript. *

      4/ The use of 4SU-seq to identify active enhancers is welcome; however, I have doubts it is working very efficiently: for instance, in the snapshots shown in Fig.2A, the very active Oct4 enhancers in ES cells are not apparent at all... More validation of the efficiency of the approach seems required.

      >Author Response

      The 4sU-seq data shown in Figure 2A was generated in samples isolated from day 3 and 4 of the ADE differentiation. It is therefore likely that the enhancers have been partly or wholly decommissioned at this point. Indeed, in a separate study we generated 4sU-seq data using the same protocol and conditions as presented here but in ES cells and differentiated NPCs (day 3 to 7) and indeed see transcription at Oct 4 enhancers in ESCs (arrowed in the screenshot shown in revision plan) which are extinguished upon differentiation to neural progenitor cells (NPCs); data from PMID: 31494034).

      5/ The effects of the EZH2 inhibitor are quite minor regarding the efficiency of the differentiation as analyzed by FACS, despite significant gene expression changes. To the knowledge of this referee, this is at odds with results obtained with Ezh2 ko ES cells that display defects in mesoderm and endoderm differentiation. I have issues reconciling these results (uncited PMID: 19026780). Either the authors perform more robust assays (inducible KOs) or they more directly explain the limitations of the study and the controversies with published work.

      >Author Response

      We agree that this result appears to be at odds with the findings in (PMID: 19026780*). This is likely due to the fact that we are acutely reducing H3K27me3 levels for a short period either during or immediately preceding the differentiation rather than removing PRC2 function genetically. This, likely provides a less pronounced defect on the ability to generate endodermal cells. However, we cannot address this without further experimentation which is beyond the scope of this study. We will more fully discuss the results in the context of this and other studies and discuss the limitations of the study in this regard. *

      Minor 1/ please add variance captured to PCA plots 2/ Fig1E add color scales to all heatmaps 3/ Fig4C,D are almost impossible to follow, please find a way to identify better the clusters/samples and make easier to correlate all the variables

      • *

      >Author Response

      *We will address all of these points in a revised manuscript. *

      Reviewer #2 (Significance (Required)):

      The paper is incremental in knowledge, and not by a big margin, as it is known already that histone mods rather channel than drive differentiation. Though, the authors do not clearly address inconsistencies with published work, especially regarding Ezh2 thought to be important to make endoderm. It is however a good addition to current knowledge, provided a better discussion of differences with published work is provided.

      >Author Response

      *As outlined above, we will address this with a more complete discussion about the distinction between the studies and what can and can’t be concluded from our approach. *

      * *

      • *

      Reviewer #3 (Evidence, reproducibility and clarity (Required)): This study investigates the role of chromatin-based regulation during cell fate specification. The authors use an ESC model of differentiation into anterior primitive streak and subsequently definitive endoderm, which they traced via a dual-reporter system that combines GSC-GFP and HHEX-RedStar. The authors mapped changes in (nascent) gene expression and histone modifications (H3K4me3/H3K27me3) at key timepoints and within different populations over six days of differentiation. Finally, the authors test the functional implications of H3K27me3 landscapes via PRC2 inhibition.

      The majority of data chart the descriptive changes in (epi)genomic and transcriptional dynamics coincident with cell differentiation. The use of nascent transcriptomics improves the temporal resolution of expression dynamics, and is an important strategy. By and large the data reinforce established paradigms. For example, that transcription is the dominant mechanism regulating mRNA levels, or that dynamic chromatin states changes occur and largely corelate of gene activity. They also identify putative enhancers with profiling data, albeit these are not validated, and confirm that PRC2 inhibition impacts cell fate processes - in this case promoting endodermal differentiation efficiency. Overall, the study is relatively well-performed and clearly written, with the omics profiling adding more datasets from in vitro cell types that can be difficult to characterise in vivo. Whilst the majority of the study may be considered incremental, the key finding is the authors conclusion that H3K27me3 is subordinate to gene activity rather than an instructive repressor. If borne out, this would mark an important observation with broad implications. However, in my view this conclusion is subject to many confounders and alternative interpretations, and the authors have not ruled out other explanations. Given the centrality of this to the novelty of the study, I would encourage further analysis/stratification of existing data, and potentially further experiments to provide more confidence in this key conclusion.

      Primary issue 1.) The authors show that at the earliest timepoint (d3), nascent gene activation of a handful of genes between G+ and G- populations is not associated with a FC loss of H3K27me3. From this the authors extrapolate their key conclusion that H3K27me3 is subordinate. Causality of chromatin modifications in gene regulation is critical to decipher, and therefore this is an important observation to confirm. Below I go through the possible confounders and issues with the conclusion at this point.

      (i) Single-cell penetrance. A possible (likely?) possibility is that gene activation initially occurs in a relatively small subset of cells at d3. Because these genes are expressed lowly prior to this, they will register as a significant upregulation in bulk analysis. However, in this scenario H3K27me3 would only be lost from a small fraction of cells, which would not be detectable against a backdrop of most cells retaining the mark. In short, the authors have not ruled out heterogeneity driving the effect. Given the different dynamic range of mRNA and chromatin marks, and that a small gain from nothing (RNA) is easier to detect than a small loss from a pre-marked state (chromatin), investigating this further is critical to draw the conclusions the authors have.

      (ii) Initial H3K27me3 levels. The plots in Fig 5 show the intersect FC of H3K27me3 and gene expression. Genes that activate at d3 show no loss of H3K27me3. However, it is important to characterise (and quantitate) whether these genes are significantly marked by H3K27me3 in the first place, which I could not find in the manuscript. Many/several of the genes may not be polycomb marked or may have low levels to begin with. This would obviously confound the analysis, since an absence/low K27 cannot be significantly lost and is unlikely to be functional. Thus, the DEG geneset should be further stratified into H3K27me3+ and K27me3- promoter groups/bins, with significance and conclusions based on the former only (e.g. boxplot in 5F).

      (iii) Sample size. The conclusions are based on a relatively small number of genes that upregulate between G+ and G- (n=55 in figure by my count, text mentions n=52). Irrespective of the other confounders above, this is quite a small subset to make the sweeping general conclusion that "loss of the repressive polycomb mark H3K27me3 is delayed relative to transcriptional activation" in the abstract. Indeed, the small number of DEG suggests the cell types being compared are similar and perhaps therefore have specific genomic features (this could be looked at) that drive .

      >Author Response

      *These are very good points and are also raised by reviewer 1 (see above). We have one example where we can definitively interrogate single cell protein expression, in our current data. Gsc (as monitored by GSC-GFP FACS and the bulk RNA analysis) meets the criteria of being robustly upregulated in all FACs sorted cells in the presence of high levels of H3K27me3 in the D3G+ population. We believe that the additional analysis (Figure R1A shown in revision plan) and the discussion above addresses the reviewer’s concerns about both the levels of expression and magnitude of H3K27me3. With respect to the third point, the numbers are low (although here I present data from the 4SU analysis with approximately three times more data points) however, the point here is not too say this happens in every instance of gene activation but more that it can happen and not just at a small subset of outlier genes. This is important, as the reviewer notes, in our understanding of how polycomb repression is relieved during development. We will also look to see if there are sequence characteristics/ motifs of these genes. In a revised manuscript we would include this data and further analysis as outlined above. The reviewer points out that the numbers vary a little between analyses. This arises due to the annotation of multiple TSSs per genes in some cases. This will be rectified throughout and made clearer in the legends. *

      Other comments: 2.) The authors show that promoter H3K4me3 corelates well with gene expression dynamics in their model. They conclude that "transcription itself is required for H3K4me3 deposition", or in other words is subordinate. This may well be the case but from their correlative data this cannot be inferred. Indeed, several recent and past papers have shown that H3K4me3 itself can directly modulate transcription, for example by triggering RNA II pause-release, by preventing epigenetic silencing and/or by recruiting the PIC. The authors could point out or discuss these alternative possibilities to provide a more balanced discourse.

      >Author Response

      We agree and this will be discussed more thoroughly and both possibilities put forward in the revised manuscript.

      3.) The labelling of some figures is unclear. In Fig 4C and 4D (right) it is impossible to tell what sample each of the lines represents. It is also not clear what the blue zone corresponds to in genome view plots (the whole gene?). Moreover, the replicate numbers are not shown in figure legends.


      >Author Response

      *We agree that the data presented in 4C and D is unclear. We will, as a minimum, collapse profiles into like populations (ESC / G- / G+ / G+H- / G+H+) which makes sense given the similarity of these populations across all analyses (see e.g. PCA analysis in Figure 1). We will also explore alternative ways of presenting the data to better highlight the dynamics and incorporate this with the changes suggested by reviewer 2. The blue shaded area represents the full extent of the key gene being discussed in the screen shot, this is mentioned in the legend but will be made clearer in a revised manuscript. Replication will also be added to the legend throughout (n=2 for ChIP-seq and n=3 for 4sU-seq). *

      4.) It would be nice to provide more discussion to reconcile the conclusions that H3K27me3 in endoderm differentiation is subordinate and the final figure showing inhibiting H3K27me3 has a significant effect on differentiation, since the latter is the functional assessment.

      >Author Response

      *We will build on the points already made that suggests that whilst K27me3 is a passive repressor that serves to act against sub-threshold activating cues, it is nonetheless a critical regulator of developmental fidelity. *

      Reviewer #3 (Significance (Required)): Overall, the study's strengths are in that it characterises epigenomic dynamics within a specific and relevant cell fate model. The nascent transcriptomics adds important resolution, and underpins the core conclusions. The weakness is that data is over-interpreted at this point, and other possibilities are not adequately tested. The conclusions should therefore either be scaled back (which reduces novelty) or further analysis and/or experiments should be performed to support the conclusion. If it proves correct, this would be a significant observation for the community,

      >Author Response

      In a revised manuscript, we will address the reviewer’s concerns with additional data and discussion as indicated above.

    1. Note: This response was posted by the corresponding author to Review Commons. The content has not been altered except for formatting.

      Learn more at Review Commons


      Reply to the reviewers

      We thank the reviewers for taking the time to read and comprehensively evaluate our manuscript. We are pleased that, overall, they recognize the quality of our data and that it supports our conclusions. We are grateful for their comments, insights and advice and have revised the manuscript accordingly as described in the point-by-point response below. We believe that the revised manuscript is substantially improved by some experimental additions, additional replicates, improved analysis and increased clarity. Some key enhancements are as follows:

      Previously we had found increased expression of the WNT pathway following CHRDL2 treatment, using RNA seq. We have now demonstrated this experimentally using the cellular levels and localisation of β-catenin. Previously we had shown that overexpression of CHRDL2 increased resistance to common chemotherapy treatments, as well as irradiation in colorectal cell lines. We have now shown that cells surviving treatment show a further reduction SMAD1/5/8 phosphorylation indicating a selection for CHRLD2 high cells during the treatment. We have also demonstrated a decrease in chemotherapy sensitivity in intestinal organoids treated with secreted forms of CHRDL2.

      1. Point-by-point description of the revisions

      This section is mandatory. *Please insert a point-by-point reply describing the revisions that were already carried out and included in the transferred manuscript. *

      Reviewer #1

      __Evidence, reproducibility and clarity __

      Clarkson and Lewis present data suggesting that overexpression of Chordin like 2 (CHRDL2) can affect colorectal cancer cell responses to chemotherapy agents, possibly by modulating stem-cell like pathways. I have the following comments:

      1. Fig. 1J-it is standard to show the images of cell migration-this is important here, given the modest effect of CHRDL2 overexpression here.

      We have now included 3 replicate control and CHRDL2 overexpressing cell images in this figure panel to support the quantification in the graph.

      Fig. 2A-the very small error bars for most of the data on the curves suggests these are n=1 experiments with multiple technical replicates to generate the error bars. Please clarify. The legend says n=3 with ANOVA analysis but no significance detected. Please clarify.

      All experiments in this figure were done with 5 technical replicates per experiment, this was replicated at least three times to give n=3 biological replicates. The error bars represent the standard error of the mean of these 3 biological replicates as stated in the legend. Some data points showed very little data variation, hence the small error bars. Raw data is available if requested.

      1. Fig. 2B-given the overlapping error bars here, how can there be a pWe have removed this representation of the data as it combined many different experiments with variable cell types and chemotherapeutics and it was difficult to carry out meaningful statistics. An overview of the data can be better seen in table form as shown in the revised figure 2B.

      Fig. 2G-did the authors try to estimate the concentration of CHRDL2 in the conditioned medium? Which cell line was used to generate this CM?

      Conditioned media was harvested from the matching transgenic cell lines with inducible CHRDL2. eg RKO cells were treated with media collected from doxycycline induced transgenic RKO cells whereas CaCO2 cells were treated with media from CaCO2 cells. The concentration of doxycycline was represented by ++ for 10ug/ml, the same notation we have used for directly induced cells treated with 10ug/ml dox.

      We did not try to quantify the absolute concentration of CHRDL2 but we have shown the relative amount on a Western blot normalised with a ponceau stain (quantification now included in supplementary figure 1).

      We have clarified our description of this experiment, inserting the following statement, "Conditioned media was harvested from corresponding cell lines with the inducible CHRDL2 transgene and the parental control cells. Induction of CHRDL2 to generate conditioned media was carried out using the same concentration and duration of doxycycline treatment as the cells in figure 2A. "

      Fig. 5-what is the potential mechanism for gene expression changes in response to CHRDL2 overexpression? Is it all due to BMP inhibition? More mechanistic detail would be welcome here.

      We have suggested other pathways involved in these functional effects based on our RNA seq data but at the moment it is not possible to say whether any changes are independent of BMP signaling. CHRDL2 is relatively understudied and as yet there is not much literature supporting BMP independent actions of CHRDL2. However, we have added some discussion and reference to an article suggesting interactions between CHRLD2 and YAP (Wang et al., 2022) including the following statement on page 17: "While the changes in BMP and WNT signaling shown in our GSEA analysis suggest that the effects of CHRDL2 in our system work directly through inhibition of BMP, it is not possible to rule out that some pathways are affected by BMP independent actions of CHRLD2. Indeed, Wang et al, suggest that CHRDL2 can directly alter phosphorylation and activity of YAP in gastric cancer cell lines, which merits further exploration (Wang et al., 2022)"

      Significance

      Unclear whether genetically engineered inducible overexpression has any real physiological significance but we all use cell models so this is OK.

      Reviewer #2

      __Evidence, reproducibility and clarity __

      Summary: In the manuscript entitled "BMP antagonist CHRDL2 enhances the cancer stem-cell phenotype and increases chemotherapy resistance in Colorectal Cancer" the authors demonstrated that Chordin-like 2 (CHDRL2), a secreted BMP antagonist, promotes a chemo-resistant colorectal cancer stem cell phenotype through the inhibition of BMP signaling. The authors took advantage of both 2D engineered colorectal cancer (CRC) cells and healthy murine 3D organoid systems. Specifically, the authors showed a decreased proliferation rate and reduced clonogenic capability upon overexpression of CHRDL2 in established human colon cancer cell lines. Subsequently, they identified a chemo-resistant phenotype upon standard therapies (5FU, Oxaliplatin and Irinotecan) in CHDRL2 overexpressing cells by performing MTS assay. The authors showed that this chemo-resistant phenotype is associated with ATM and RAD21 activation, supporting an induction of DNA damage signaling pathway. Of note, the authors assessed that the exposure of 3D murine organoid to CHRDL2 resulted in a stem-like phenotype induction accompanied by a reduction of the differentiated counterpart. From RNA-seq data analysis emerged the upregulation of genes associated to stemness and DNA repair pathways in CHRDL2 overexpressing cells.

      Major comments: 1. In the first paragraph of the result section authors assessed that "Colorectal adenocarcinoma cell lines were deliberately chosen to encompass a range of CHRDL2 expression levels and genetic mutations", without showing qRT-PCR or WB data on the differential expression levels of CHRDL2 in a panel of immortalized CRC cell lines. Authors should include these data to better support their choice.

      *We have now included some qRT-PCR in supplementary figure 1 alongside a table of some of the key driver mutations in each cell line. Western blotting of these cells shows only a very low concentration of CHRDL2 protein. As shown in figure 1B in the control columns, no significant protein expression is observed in any line. *

      In Figure 1F, authors described a reduction of cell proliferation in CRC cell lines expressing high levels of CHRDL2 only under low glucose conditions. Why did the authors perform the assay under these conditions? They should better argue this aspect and validated the role of CHRDL2 in metabolism rewiring by performing additional in vitro assays.

      We have removed this aspect of the paper as it does not add significantly to our overall conclusions and we can clearly see the effects of CHRDL2 overexpression under standard growth conditions (Figure 1G).

      The authors should evaluate the role of CHRDL2 in promoting a stem-like phenotype in human colon cancer stem cells freshly isolated from patients and characterized.

      We would very much like to do experiments such as this but it is beyond the scope of this study and will be included in upcoming grant proposals.

      In order to confirm the data obtained on 3D murine organoids system obtained from normal Intestinal Stem Cells, authors should investigate the stemness induction, driven by CHRDL2, also in human intestinal organoids.

      Experiments using human intestinal organoids are currently planned and ethical approval applications and grant proposals are underway for future experiments of this nature.

      The authors should evaluate the oncogenic role of CHRDL2, through the maintenance of stemness, by performing orthotopic or subcutaneous experiments in vivo model.

      Similarly, this is not possible for this manuscript but is planned for the future alongside a transgenic mouse model of inducible CHRDL2 overexpression in the intestine.

      BMPs proteins are part of a very broad protein family. In the introduction section, authors should indicate the specific BMP protein on which CHRDL2 exerts its inhibitory function. Moreover, they should have assessed BMP protein levels in CACO2, LS180, COLO320 and RKO cell lines.

      We have clarified the interactions between CHRDL2 and specific BMPs in the introduction. We have not specifically assessed the BMP protein levels in our cells however we have now included an analysis of expression data from the Cancer Cell Line Encyclopedia in supplementary figure 1 C.

      In first panel, the authors should quantify the secreted levels of CHRDL2 in the media of overexpressing CHRDL2 cell lines.

      We have done this using the ponceau staining as a loading control and the results are displayed (supplementary figure 1).

      In Figure 2D the authors should use the appropriate controls and describe this with more details in results section.

      In this figure we have used Hoechst staining followed by FACs analysis to identify the cell cycle profile of our CHRDL2 treated cells. We have improved the description of this in the methods section. Appropriate controls for staining, both negative and positive, are used when setting up the analysis for this experiment. The cell cycle profile is calculated using the Novocyte in house software. We have now included the histogram plots in the main figure to clarify these data in figure 2D.

      In Figure 3A, the authors should have performed the assay by choosing IC50.

      *We attempted these experiments with the IC50 levels, however the high amount of cell death and frequency of apoptotic cells meant that clear images were difficult to obtain. We therefore reduced the concentrations and still had very measurable effects. *

      In Supplementary Fig. 4A-B. the results are unclear. The control cell lines are already chemo resistant.

      Again, we used IC25 levels of the drugs so that our cells were damaged but still live throughout the experiment. This has been explained on page 10.

      The authors should review and add statistical analysis in both main and supplementary figures.

      *We have now added additional details about statistical analysis throughout the figures, legend and main text, showing all significance levels as well as non-significance for each data set. * Minor comments: 1. The quality of immunofluorescence and WB images should be implemented, and in the immunofluorescence panels scale bars should be added.

      We have added or improved scale bars on each immunofluorescence image. Western blot images have been improved.

      In the graphical abstract authors reported that CHRDL2 overexpression increase WNT and EMT pathways, without performing any specific assay to demonstrate this. Authors should correct and graphically improve the graphical abstract.

      *This is a good point and we have now carried out Beta-catenin immunofluorescence as a measure of WNT signaling on both our cancer cell lines - showing an increase in nuclear beta-catenin (figure 1J and K), and our organoids - showing an increase in overall levels and cytoplasmic staining (Figure 4 F). In terms of EMT markers we have carried out immunofluorescence on IQGAP1 (Figure 1I). IQGAP1 is significantly upregulated in CHRDL2 cells, reflecting its role in reduced cell adhesion and increased migration. This correlates with our data showing increased cellular migration as well as the increase in EMT related transcription in our RNAseq data. *

      The term "significantly" in the discussion section is inappropriately referred to data showed in the histogram in Figure 1J. Moreover, in Figure 1Jthe authors should delete from the y-axis the term "corrected".

      We have changed significantly to substantially

      The term "significant" in discussion is inappropriately referred to BMI1 expression level if compared to the histogram in Figure 4G.

      We have changed significantly to "a trend to increase"

      In Figure 2C the authors should add the unit of measurement (fold over control) in the table.

      We have done this

      In Figure 4E the authors should add the figure legend reporting OLFM4 protein.

      We have done this

      The authors should include few sentences summarizing the findings at the end of each paragraph.

      We have added short summaries at the start or end of each section to improve the flow of the results section.

      Significance

      General assessment: Overall, the work is aimed to elucidate the role of CHRDL2 already considered a poor prognosis biomarker involved in the promotion of CRC (PMID: 28009989), in promoting stem-like properties. The authors elucidated new additional insights into the molecular mechanisms regulating stemness phenotype induced by the BMP antagonist CHRDL2 in CRC. The authors include in the study a large amount of data, which only partially support their hypothesis. However, this manuscript lacks organization and coherence, making it challenging to follow and read. Numerous concerns need to be addressed, along with some sentences to rephrase in the result and discussion sections.

      Advance: The manuscript reported some functional insights on the role of CHRDL2 in colorectal cancer, but additional data should be added to support authors 'conclusions.

      Audience: The manuscript is suggested for basic research scientists.

      __Reviewer #3 (Evidence, reproducibility and clarity (Required)):____ __ Summary BMP antagonist CHRDL2 enhances the cancer stem-cell phenotype and increases chemotherapy resistance in Colorectal Cancer Eloise Clarkson et al. The manuscript explored the function of CHRDL2, a BMP antagonist, on colorectal cancer (CRC). The authors found that CHRDL2 overexpression can enhance the survival of CRC cells during chemotherapy and irradiation treatment with elevated levels of stem-cell markers and reduced differentiation. Further RNA-seq analysis revealed that CHRDL2 increased the expression of stem-cell markers, WNT signaling and other well-established cancer-associated pathways. Overall, the manuscript is well-written and presented. I have some suggestions:

      Major points:

      1. The authors assert BMP antagonism was demonstrated by assessing levels of phosphorylated SMAD1/5 (Figure 1G). However, the immunoblotting assay only depicted P-SMAD 1/5 levels and B-ACTIN as internal control. It's suggested to include total-SMAD1/5 immunoblotting as an internal control to further support the claim of BMP antagonism.

      The reviewer is correct that this is the best control. Western blotting has now been performed with total SMAD1 protein expression used as an internal control and this is shown in Figure 1D and Supplementary figure 1F

      The authors argue that CHRDL2 overexpression reduced the proliferation of CRC cell lines, as evidenced by cell proliferation assays. However, from Figure 1E, the reduction in proliferation appears insignificant. It would be beneficial to perform one-way ANOVA tests on each time point for CHRDL2+ and CHRDL2++ with Control in Figure 1E to ascertain significance.

      *We now have repeated this experiment to reduce variability and have also provided two-way ANOVA analysis between Control and CHRDL2+ and Control and CHRDL2++. One-way ANOVA at timepoint 96hr also provided with details in the figure legend. *

      The findings indicating that overexpressing CHRDL2 can confer resistance to chemotherapy in CRC cells (Figure 2A-C) are noteworthy. To deepen the understanding of BMP signaling in cancer stemness and the molecular underpinning of CHRDL2 antagonism, additional western blot assays on P-SMAD1/5 with CHRDL2 overexpression and drug treatment are recommended.

      *Western blotting of P-SMAD1/5 upon cells treated with IC50 5FU has now been performed in figure 2C (in the same experiment as the revised panels in figure 1D). The data suggest that CHRDL2 overexpressing cells able to survive chemotherapy have higher levels of P-SMAD1/5 reduction compared to that of untreated cells, strongly suggesting that chemotherapy treatment acts to select the cells with the highest CHRDL2 expression. We thank reviewer 3 for this suggested experiment and have included further discussion on this on page 8. *

      The assertion that extrinsic CHRDL2 addition diminishes differentiation and enhances stem-cell numbers in an intestinal organoid model is intriguing. As BMP signaling inhibition contributes to intestinal cell stemness, incorporating additional layers for BMP antagonism of CHRDL2 on intestinal organoids through immunoblotting or real-time quantitative PCR for treated organoids would augment the conclusions.

      As stated in the response to reviewer 2, we have investigated Beta-catenin in our organoids following CHRDL2 treatment using immunofluorescence and find that the levels are increased with the staining shifting from the membrane to the cytoplasm and nucleus (Figure 4F).

      The authors claim CHRDL2 overexpression can decrease BMP signaling based on GSEA analysis (Figure 5E). However, the GSEA results did not demonstrate the downregulation of BMP signaling. Reanalysis of this GSEA analysis is warranted.

      *We agree with this point and have changed the description of this result since the gene set covers both positive and negative regulators of the BMP pathway. We cannot conclusively say from this RNAseq data set that BMP signaling is "downregulated", however since SMAD phosphorylation is increased and nuclear beta-catenin is increased, overall we suggest that the changes we see are likely to represent the effects of decreased BMP signaling along with increased WNT signaling. *

      Minor Points:

      6.Provide the threshold/cutoff values chosen for differential expressed genes (DEGs) in CHRDL2+ and CHRDL2++ RNA-seq compared with control cells. Explain the minimal overlap between CHRDL2 LOW and CHRDL2 HIGH DEGs. Consider presenting all DEGs in CHRDL2 LOW and CHRDL2 HIGH compared with control cells in one gene expression heatmap for better visualization.

      We have now provided the cutoff values for the DEGs in the legend for figure 5 (PThe minimal overlap of DEGs in the low and high expressing cells is an interesting point. We hypothesize that this may be related to the different effects of intermediate vs high levels of WNT signaling that occurs in colon cancer cells, frequently discussed in the literature as the "Just right hypothesis" (Lamlum et al. 1999, Albuquerque et al., 2002, Lewis et al., 2010). However, we haven't included this in the discussion as it merits further exploration. However, we have mainly focused on specific genes that are modified in both data sets, which are more likely to be the direct result of CHRDL2 modification. *

      After DEGs analysis, perform Gene Ontology (GO) analysis on these DEGs to further investigate possible gene functions rather than selectively discussing some genes, enhancing understanding of CHRDL2 functions in CRC cells.

      We have carried out this analysis using a variety of tools and have now included a Gene Ontology Panther analysis as supplementary figure 7. We have included a comment on this in the text on page 14 saying "Gene ontology analysis supports these findings with enrichment in biological processes such as cellular adhesion, apoptosis and differentiation. "

      Conduct similar experiments in both 2D culture and organoid systems, if feasible, to provide more comprehensive insights into CHRDL2's oncogenic roles in CRC tumor progression.

      *We have now performed chemotherapy treatment on our organoid systems, and have found that organoids with extrinsic CHRDL2 addition have a higher survival rate after chemotherapy compared to a control (Figure 4H and I). *

      Label significance (*, **, ***, and n.s.) for every CRC cell line treated with CHRDL2 in Figure 2D, 2F, 2J, 4G, 5D, and 5F.

      We have done this

      Label the antibodies with different colors used for immunofluorescence in the figure text in Figure 4E.

      We have done this

      * * Include replicate dots for the Control group in the bar plots in Figure 1F and 2B.

      We have done this

      * * Add scale bars in Figure 3A and correct similar issues in other figures if applicable.

      We have done this

      * *13.Correct grammar and punctuation mistakes throughout the manuscript. For example:

      We have done this and further proofread our revised manuscript

      Page 7: "As seen in Figure 1J, CHRDL2 overexpression significantly increased the number of migrated cells (P *We have now added additional details about statistical analysis throughout the figures, legend and main text, showing all significance levels as well as non-significance for each data set. * Reviewer #3 (Significance (Required)):

      The current study presents compelling evidence demonstrating that BMP signaling antagonist CHRDL2 enhances colon stem cell survival in colorectal cancer cell lines and organoid models. Further validation through CRC mouse models could offer invaluable insights into the clinical relevance and therapeutic implications of CHRDL2 in colorectal cancer.

    1. Reviewer #1 (Public Review):

      The manuscript by Wang et al is, like its companion paper, very unusual in the opinion of this reviewer. It builds off of the companion theory paper's exploration of the "Wright-Fisher Haldane" model but applies it to the specific problem of diversity in ribosomal RNA arrays. The authors argue that polymorphism and divergence among rRNA arrays are inconsistent with neutral evolution, primarily stating that the amount of polymorphism suggests a high effective size and thus a slow fixation rate, while we, in fact, observe relatively fast fixation between species, even in putatively non-functional regions. They frame this as a paradox in need of solving, and invoke the WFH model.

      The same critiques apply to this paper as to the presentation of the WFH model and the lack of engagement with the literature, particularly concerning Cannings models and non-diffusive limits. However, I have additional concerns about this manuscript, which I found particularly difficult to follow.

      My first, and most major, concern is that I can never tell when the authors are referring to diversity in a single copy of an rRNA gene compared to when they are discussing diversity across the entire array of rRNA genes. I admit that I am not at all an expert in studies of rRNA diversity, so perhaps this is a standard understanding in the field, but in order for this manuscript to be read and understood by a larger number of people, these issues must be clarified.

      The authors frame the number of rRNA genes as roughly equivalent to expanding the population size, but this seems to be wrong: the way that a mutation can spread among rRNA gene copies is fundamentally different than how mutations spread within a single copy gene. In particular, a mutation in a single copy gene can spread through vertical transmission, but a mutation spreading from one copy to another is fundamentally horizontal: it has to occur because some molecular mechanism, such as slippage, gene conversion, or recombination resulted in its spread to another copy. Moreover, by collapsing diversity across genes in an rRNA array, the authors are massively increasing the mutational target size.

      For example, it's difficult for me to tell if the discussion of heterozygosity at rRNA genes in mice starting on line 277 is collapsed or not. The authors point out that Hs per kb is ~5x larger in rRNA than the rest of the genome, but I can't tell based on the authors' description if this is diversity per single copy locus or after collapsing loci together. If it's the first one, I have concerns about diversity estimation in highly repetitive regions that would need to be addressed, and if it's the second one, an elevated rate of polymorphism is not surprising, because the mutational target size is in fact significantly larger.

      Even if these issues were sorted out, I'm not sure that the authors framing, in terms of variance in reproductive success is a useful way to understand what is going on in rRNA arrays. The authors explicitly highlight homogenizing forces such as gene conversion and replication slippage but then seem to just want to incorporate those as accounting for variance in reproductive success. However, don't we usually want to dissect these things in terms of their underlying mechanism? Why build a model based on variance in reproductive success when you could instead explicitly model these homogenizing processes? That seems more informative about the mechanism, and it would also serve significantly better as a null model, since the parameters would be able to be related to in vitro or in vivo measurements of the rates of slippage, gene conversion, etc.

      In the end, I find the paper in its current state somewhat difficult to review in more detail, because I have a hard time understanding some of the more technical aspects of the manuscript while so confused about high-level features of the manuscript. I think that a revision would need to be substantially clarified in the ways I highlighted above.

    1. Author response:

      The following is the authors’ response to the previous reviews.

      Public Reviews: 

      Reviewer #1 (Public Review): 

      Summary: 

      This paper uses a model of binge alcohol consumption in mice to examine how the behaviour and its control by a pathway between the anterior insular cortex (AIC) to the dorsolateral striatum (DLS) may differ between males and females. Photometry is used to measure the activity of AIC terminals in the DLS when animals are drinking and this activity seems to correspond to drink bouts in males but not females. The effects appear to be lateralized with inputs to the left DLS being of particular interest. 

      Strengths: 

      Increasing alcohol intake in females is of concern and the consequences for substance use disorder and brain health are not fully understood, so this is an area that needs further study. The attempt to link fine-grained drinking behaviour with neural activity has the potential to enrich our understanding of the neural basis of behaviour, beyond what can be gleaned from coarser measures of volumes consumed etc. 

      Weaknesses: 

      The introduction to the drinking in the dark (DID) paradigm is rather narrow in scope (starting line 47). This would be improved if the authors framed this in the context of other common intermittent access paradigms and gave due credit to important studies and authors that were responsible for the innovation in this area (particularly studies by Wise, 1973 and returned to popular use by Simms et al 2010 and related papers; e.g., Wise RA (1973). Voluntary ethanol intake in rats following exposure to ethanol on various schedules. Psychopharmacologia 29: 203-210; Simms, J., Bito-Onon, J., Chatterjee, S. et al. Long-Evans Rats Acquire Operant Self-Administration of 20% Ethanol Without Sucrose Fading. Neuropsychopharmacol 35, 1453-1463 (2010).)

      We appreciate the reviewer’s perspective on the history of the alcohol research field. There are hundreds of papers that could be cited regarding all the numerous different permutations of alcohol drinking paradigms. This study is an eLife “Research Advances” manuscript that is a direct follow-up study to a previously published study in eLife (Haggerty et al., 2022) that focused on the Drinking in the Dark model of binge alcohol drinking. This study must be considered in the context of that previous study (they are linked), and thus we feel that a comprehensive review of the literature is not appropriate for this study.

      The original drinking in the dark demonstrations should also be referenced (Rhodes et al., 2005). Line 154 Theile & Navarro 2014 is a review and not the original demonstration. 

      This is a good recommendation. We have added this citation to Line 33 and changed Line 154.

      When sex differences in alcohol intake are described, more care should be taken to be clear about whether this is in terms of volume (e.g. ml) or blood alcohol levels (BAC, or at least g/kg as a proxy measure). This distinction was often lost when lick responses were being considered. If licking is similar (assuming a single lick from a male and female brings in a similar volume?), this might mean males and females consume similar volumes, but females due to their smaller size would become more intoxicated so the implications of these details need far closer consideration. What is described as identical in one measure, is not in another. 

      As shown in Figure 1, all measures of intake are reported as g/kg for both water and alcohol to assess intakes across fluids that are controlled by body weights. We do not reference changes in fluid volume or BACs to compare differences in measured lickometry or photometric signals, except in one instance where we suggest that the total volume of water (ml) is greater than the total amount of alcohol (ml) consumed in DID sessions, but this applies generally to all animals, regardless of sex, across all the experimental procedures.

      In Figure 2 – Figure Supplement 1 we show drinking microstructures across single DID sessions, and that males and females drink similarly, but not identically, when assessing drinking measures at the smallest timescale that we have the power to detect with the hardware we used for these experiments. Admittedly, the variability seen in these measures is certainly non-zero, and while we are tempted to assume that there exist at least some singular drinks that occur identically between males and females in the dataset that support the idea that females are simply just consuming more volume of fluid per singular drink, we don’t have the sampling resolution to support that claim statistically. Further, even if females did consume more volume per singular drink that males, we do not believe that is enough information to make the claim that such behavior leads to more “intoxication” in females compared males, as we know that alcohol behaviors, metabolism, and uptake/clearance all differ significantly by sex and are contributing factors towards defining an intoxication state. We’ve amended the manuscript to remove any language of referencing these drinking behaviors as identical to clear up the language.

      No conclusions regarding the photometry results can be drawn based on the histology provided. Localization and quantification of viral expression are required at a minimum to verify the efficacy of the dual virus approach (the panel in Supplementary Figure 1 is very small and doesn't allow terminals to be seen, and there is no quantification). Whether these might differ by sex is also necessary before we can be confident about any sex differences in neural activity. 

      We provide hit maps of our fiber placements and viral injection centers, as we have, and many other investigators do regularly for publication based on histological verification. Figure 1A clearly shows the viral strategy taken to label AIC to DLS projections with GCaMP7s, and a representative image shows green GCaMP positive terminals below the fiber placement. Considering the experiments, animals without proper viral expression did not display or had very little GCaMP signal, which also serves as an additional expression-based control in addition to typical histology performed to confirm “hits”. These animals with poor expression or obvious misplacement of the fiber probes were removed as described in the methods. Further, we also report our calcium signals as z-scored differences in changes in observed fluorescence, thus we are comparing scaled averages of signals across sexes, and days, which helps minimize any differences between “low” or “high” viral transduction levels at the terminals, directly underneath the tips of the fibers.

      While the authors have some previous data on the AIC to DLS pathway, there are many brain regions and pathways impacted by alcohol and so the focus on this one in particular was not strongly justified. Since photometry is really an observational method, it's important to note that no causal link between activity in the pathway and drinking has been established here. 

      As mentioned above, this article is an eLife Research Advances article that builds on our previous AIC to DLS work published in eLife (Haggerty et al., 2022). Considering that this is a linked article, a justification for why this brain pathway was chosen is superfluous. In addition, an exhaustive review of all the different brain regions and pathways that are affected by binge alcohol consumption to justify this pathway seems more appropriate to a review article than an article such as this.  

      We make no claims that photometric recordings are anything but observational, but we did observe these signals to be different when time-locked to the beginning of drinking behaviors. We describe this link between activity in the pathway and drinking throughout the manuscript. It is indeed correlational, but just because it is not causal does not mean that our findings are invalid or unimportant.

      It would be helpful if the authors could further explain whether their modified lickometers actually measure individual licks. While in some systems contact with the tongue closes a circuit which is recorded, the interruption of a photobeam was used here. It's not clear to me whether the nose close to the spout would be sufficient to interrupt that beam, or whether a tongue protrusion is required. This detail is important for understanding how the photometry data is linked to behaviour. The temporal resolution of the GCaMP signal is likely not good enough to capture individual links but I think more caution or detail in the discussion of the correspondence of these events is required. 

      The lickometers do not capture individual licks, but a robust quantification of the information they capture is described in Godynyuk et al. 2019 and referenced in multiple other papers (Flanigan et al. 2023, Haggerty et al. 2022, Grecco et al. 2022, Holloway et al. 2023) where these lickometers have been used. However, individual lick tracking is not a requirement for tracking drinking behaviors more generally. The lickometers used clearly track when the animals are at the bottles, drinking fluids, and we have used the start of that lickometer signal to time-lock our photometry signals to drinking behaviors. We make no claims or have any data on how photometric signals may be altered on timescales of single licks. In regard to how AIC to DLS signals change on the second time scale when animals initiate drinking behaviors, we believe we explain these signals with caution and in context of the behaviors they aim to describe.

      Even if the pattern of drinking differs between males and females, the use of the word "strategy" implies a cognitive process that was never described or measured. 

      We use the word strategy to describe a plan of action that is executed by some chunking of motor sequences that amounts to a behavioral event, in this case drinking a fluid. We do not mean to imply anything further than this by using this specific word.

      Reviewer #2 (Public Review): 

      Summary: 

      This study looks at sex differences in alcohol drinking behaviour in a well-validated model of binge drinking. They provide a comprehensive analysis of drinking behaviour within and between sessions for males and females, as well as looking at the calcium dynamics in neurons projecting from the anterior insula cortex to the dorsolateral striatum. 

      Strengths: 

      Examining specific sex differences in drinking behaviour is important. This research question is currently a major focus for preclinical researchers looking at substance use. Although we have made a lot of progress over the last few years, there is still a lot that is not understood about sex-differences in alcohol consumption and the clinical implications of this. 

      Identifying the lateralisation of activity is novel, and has fundamental importance for researchers investigating functional anatomy underlying alcohol-driven behaviour (and other reward-driven behaviours). 

      Weaknesses: 

      Very small and unequal sample sizes, especially females (9 males, 5 females). This is probably ok for the calcium imaging, especially with the G-power figures provided, however, I would be cautious with the outcomes of the drinking behaviour, which can be quite variable. 

      For female drinking behaviour, rather than this being labelled "more efficient", could this just be that female mice (being substantially smaller than male mice) just don't need to consume as much liquid to reach the same g/kg. In which case, the interpretation might not be so much that females are more efficient, as that mice are very good at titrating their intake to achieve the desired dose of alcohol. 

      We agree that the “more efficient” drinking language could be bolstered by additional discussion in the text, and thus have added this to the manuscript starting at line 440.

      I may be mistaken, but is ANCOVA, with sex as the covariate, the appropriate way to test for sex differences? My understanding was that with an ANCOVA, the covariate is a continuous variable that you are controlling for, not looking for differences in. In that regard, given that sex is not continuous, can it be used as a covariate? I note that in the results, sex is defined as the "grouping variable" rather than the covariate. The analysis strategy should be clarified. 

      In lines 265-267, we explicitly state that the covariate factor was sex, which is mathematically correct based on the analyses we ran. We made an in-text error where we referred to sex as a grouping variable on Line 352, when it should have been the covariate. Thank you for the catch and we have corrected the manuscript.

      But, to reiterate, we are attempting to determine if the regression fits by sex are significantly different, which would be reported as a significant covariate. Sex is certainly a categorical variable, but the two measures at which we are comparing them against are continuous, so we believe we have the validity to run an ANCOVA here.

      Reviewer #3 (Public Review): 

      Summary: 

      In this manuscript by Haggerty and Atwood, the authors use a repeated binge drinking paradigm to assess how water and ethanol intake changes in male in female mice as well as measure changes in anterior insular cortex to dorsolateral striatum terminal activity using fiber photometry. They find that overall, males and females have similar overall water and ethanol intake, but females appear to be more efficient alcohol drinkers. Using fiber photometry, they show that the anterior insular cortex (AIC) to dorsolateral striatum projections (DLS) projections have sex, fluid, and lateralization differences. The male left circuit was most robust when aligned to ethanol drinking, and water was somewhat less robust. Male right, and female and left and right, had essentially no change in photometry activity. To some degree, the changes in terminal activity appear to be related to fluid exposure over time, as well as within-session differences in trial-by-trial intake. Overall, the authors provide an exhaustive analysis of the behavioral and photometric data, thus providing the scientific community with a rich information set to continue to study this interesting circuit. However, although the analysis is impressive, there are a few inconsistencies regarding specific measures (e.g., AUC, duration of licking) that do not quite fit together across analytic domains. This does not reduce the rigor of the work, but it does somewhat limit the interpretability of the data, at least within the scope of this single manuscript. 

      Strengths: 

      - The authors use high-resolution licking data to characterize ingestive behaviors. 

      - The authors account for a variety of important variables, such as fluid type, brain lateralization, and sex. 

      - The authors provide a nice discussion on how this data fits with other data, both from their laboratory and others'. 

      - The lateralization discovery is particularly novel. 

      Weaknesses: 

      - The volume of data and number of variables provided makes it difficult to find a cohesive link between data sets. This limits interpretability.

      We agree there is a lot of data and variables within the study design, but also believe it is important to display the null and positive findings with each other to describe the changes we measured wholistically across water and alcohol drinking.

      - The authors describe a clear sex difference in the photometry circuit activity. However, I am curious about whether female mice that drink more similarly to males (e.g., less efficiently?) also show increased activity in the left circuit, similar to males. Oppositely, do very efficient males show weaker calcium activity in the circuit? Ultimately, I am curious about how the circuit activity maps to the behaviors described in Figures 1 and 2. 

      In Figure 3C, we show that across the time window of drinking behaviors, that female mice who drink alcohol do have a higher baseline calcium activity compared to water drinking female mice, so we believe there are certainly alcohol induced changes in AIC to DLS within females, but there remains to be a lack of engagement (as measured by changes in amplitude) compared to males. So, when comparing consummatory patterns that are similar by sex, we still see the lack of calcium signaling near the drinking bouts, but small shifts in baseline activity that we aren’t truly powered to resolve (using an AUC or similar measurements for quantification) because the shifts are so small. Ultimately, we presume that the AIC to DLS inputs in females aren’t the primary node for encoding this behavior, and some recent work out of David Werner’s group (Towner et al. 2023) suggests that for males who drink, the AIC becomes a primary node of control, whereas in females, the PFC and ACC, are more engaged. Thus, the mapping of the circuit activity onto the drinking behaviors more generally represented in Figures 1 and 2 may be sexually dimorphic and further studies will be needed to resolve how females engage differential circuitry to encode ongoing binge drinking behaviors.

      - What does the change in water-drinking calcium imaging across time in males mean? Especially considering that alcohol-related signals do not seem to change much over time, I am not sure what it means to have water drinking change. 

      The AIC seems to encode many physiologically relevant, interoceptive signals, and the water drinking in males was also puzzling to us as well. Currently, we think it may be both the animals becoming more efficient at drinking out of the lickometers in early weeks and may also be signaling changes due to thirst states of taste associated with the fluid. While this is speculation, we need to perform more in-depth studies to determine how thirst states or taste may modulate AIC to DLS inputs, but we believe that is beyond the scope of this current study.

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors): 

      Line 45 - states alcohol use rates are increasing in females across the past half-decade. I thought this trend was apparent over the past half-century? Please consider revising this. 

      According to NIAAA, the rates of alcohol consumption in females compares to males has been closing for about the past 100 years now, but only recently are those trends starting to reverse, where females are drinking similar amounts or more than males.

      Placing more of the null findings into supplemental data would make the long paper more accessible to the reader. 

      In reference to reviewer’s three’s point as well, there is a lot of data we present, and we hope for others to use this data, both null and positive findings in their future work. As formatted on eLife’s website, we think it is important to place these findings in-line as well.

      Reviewer #2 (Recommendations For The Authors): 

      In addition to the points raised about analysis and interpretation in the Public Review, I have a minor concern about the written content. I find the final sentence of the introduction "together these findings represent targets for future pharmacotherapies.." a bit unjustified and meaningless. The findings are important for a basic understanding of alcohol drinking behaviour, but it's unclear how pharmacotherapies could target lateralised aic inputs into dls. 

      There are on-going studies (CANON-Pilot Study, BRAVE Lab, Stanford) for targeted therapies that use technologies like TMS and focused ultrasound to activate the AIC to alleviate alcohol cravings and decrease heavy drinking days. The difficulty with these next-generation therapeutics is often targeting, and thus we think this work may be of use to those in the clinic to further develop these treatments. We agree that this data does not support the development of pharmacotherapies in a traditional sense, and thus have removed the word and added text to reference TMS and ultrasound approaches to bolster this statement in lines 101+.

    1. Rather than assuming that the entrepreneurialpersonality can be characterized by set of unified traits, Steyaert (2007a) contends thatthe question should be approached from a narrative point of view

      I really like this angle (and it's why I highlighted the "narrative turn" on pg. 230. This article engages in a "selective reading" (I love that term) of Branson's narrative (literally, his autobiography). But every account of entrepreneurs and their successes (and failures/struggles to succeed) can be understood as a story. We're not just reading the story of that individual (though that's often the focus); we are also encountering the story of entrepreneurship and the process of its ongoing fashioning/re-imagining. When I ask you to consider how you're entrepreneurial, I'm asking you to re-write your own story...

    2. create structures of desire that teach us how to desireto become an entrepreneur

      all of the success stories surrounding entrepreneurs encourage us to want to be entrepreneurs... and not just the heroic version either! All the narratives of people with side-hustles teach us to desire a similar pursuit, selling a gig-economy and precarious labour under the guise of passion and self-improvement... It's not just that we are all entrepreneurs (Szeman), but that we desire to be... (because we all DESIRE (or ought to desire, supposedly) more success, more freedom, more independence (financially and otherwise)...)

      And maybe, the more we desire and dream, the more we might do (converting our desires into action and making our dreams reality !!)

    3. the desire for transgression (overcoming oneself) and the desire forauthenticity (becoming oneself) make up the entrepreneurial subjectivity

      not just in Branson's case either. I think it's reasonable to suggest that these are potentially pillars in any entrepreneurial mindset.

    1. Rather than assuming that the entrepreneurialpersonality can be characterized by set of unified traits, Steyaert (2007a) contends thatthe question should be approached from a narrative point of view

      I really like this angle (and it's why I highlighted the "narrative turn" on pg. 230. This article engages in a "selective reading" (I love that term) of Branson's narrative (literally, his autobiography). But every account of entrepreneurs and their successes (and failures/struggles to succeed) can be understood as a story. We're not just reading the story of that individual (though that's often the focus); we are also encountering the story of entrepreneurship and the process of its ongoing fashioning/re-imagining. When I ask you to consider how you're entrepreneurial, I'm asking you to re-write your own story...

    2. create structures of desire that teach us how to desireto become an entrepreneur

      all of the success stories surrounding entrepreneurs encourage us to want to be entrepreneurs... and not just the heroic version either! All the narratives of people with side-hustles teach us to desire a similar pursuit, selling a gig-economy and precarious labour under the guise of passion and self-improvement... It's not just that we are all entrepreneurs (Szeman), but that we desire to be... (because we all DESIRE (or ought to desire, supposedly) more success, more freedom, more independence (financially and otherwise)...)

      And maybe, the more we desire and dream, the more we might do (converting our desires into action and making our dreams reality !!)

    3. the desire for transgression (overcoming oneself) and the desire forauthenticity (becoming oneself) make up the entrepreneurial subjectivity

      not just in Branson's case either. I think it's reasonable to suggest that these are potentially pillars in any entrepreneurial mindset.

    1. Author response:

      eLife assessment

      The authors present a potentially useful approach of broad interest arguing that anterior cingulate cortex (ACC) tracks option values in decisions involving delayed rewards. The authors introduce the idea of a resource-based cognitive effort signal in ACC ensembles and link ACC theta oscillations to a resistance-based strategy. The evidence supporting these new ideas is incomplete and would benefit from additional detail and more rigorous analyses and computational methods.

      The reviewers have provided several excellent suggestions and pointed out important shortcomings of our manuscript. We are grateful for their efforts. To address these concerns, we are planning a major revision to the manuscript. In the revision, our goal is to address each of the reviewer’s concerns and codify the evidence for resistance- and resource-based control signals in the rat anterior cingulate cortex. We have provided a nonexhaustive list we plan to address in the point by point responses below.   

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      Young (2.5 mo [adolescent]) rats were tasked to either press one lever for immediate reward or another for delayed reward.

      Please note that at the time of testing and training that the rats were > 4 months old.

      The task had a complex structure in which (1) the number of pellets provided on the immediate reward lever changed as a function of the decisions made, (2) rats were prevented from pressing the same lever three times in a row. Importantly, this task is very different from most intertemporal choice tasks which adjust delay (to the delayed lever), whereas this task held the delay constant and adjusted the number of 20 mg sucrose pellets provided on the immediate value lever.

      Several studies parametrically vary the immediate lever (PMID: 39119916, 31654652, 28000083, 26779747, 12270518, 19389183). While most versions of the task will yield qualitatively similar estimates of discounting, the adjusting amount is preferred as it provides the most consistent estimates (PMID: 22445576). More specifically this version of the task avoids contrast effects of that result from changing the delay during the session (PMID: 23963529, 24780379, 19730365, 35661751) which complicates value estimates.

      Analyses are based on separating sessions into groups, but group membership includes arbitrary requirements and many sessions have been dropped from the analyses.

      We are in discussions about how to address this valid concern. This includes simply splitting the data by delay. This approach, however, has conceptual problems that we will also lay out in a full revision.  

      Computational modeling is based on an overly simple reinforcement learning model, as evidenced by fit parameters pegging to the extremes.

      We apologize for not doing a better job of explaining the advantages of this type of model for the present purposes. Nevertheless, given the clear lack of enthusiasm, we felt it was better to simply update the model as suggested by the Reviewers. The straightforward modifications have now been implemented and we are currently in discussion about how the new results fit into the larger narrative.

      The neural analysis is overly complex and does not contain the necessary statistics to assess the validity of their claims.

      We plan to streamline the existing analysis and add statistics, where required, to address this concern.

      Strengths:

      The task is interesting.

      Thank you for the positive comment

      Weaknesses:

      Behavior:

      The basic behavioral results from this task are not presented. For example, "each recording session consisted of 40 choice trials or 45 minutes". What was the distribution of choices over sessions? Did that change between rats? Did that change between delays? Were there any sequence effects? (I recommend looking at reaction times.) Were there any effects of pressing a lever twice vs after a forced trial?

      Animals tend to make more immediate choices as the delay is extended, which is reflected in Figure 1. We will add more detail and additional statistics to address these questions. 

      This task has a very complicated sequential structure that I think I would be hard pressed to follow if I were performing this task.

      Human tasks implement a similar task structure (PMID: 26779747). Please note the response above that outlines the benefits of using of this task.   

      Before diving into the complex analyses assuming reinforcement learning paradigms or cognitive control, I would have liked to have understood the basic behaviors the rats were taking. For example, what was the typical rate of lever pressing? If the rats are pressing 40 times in 45 minutes, does waiting 8s make a large difference?

      This is a good suggestion. However, rats do not like waiting for rewards, even small delays. Going from the 4 à 8 sec delay results in more immediate choices, indicating that the rats will forgo waiting for a smaller reinforcer at the 8 sec delay as compared to the 4 sec.  

      For that matter, the reaction time from lever appearance to lever pressing would be very interesting (and important). Are they making a choice as soon as the levers appear? Are they leaning towards the delay side, but then give in and choose the immediate lever? What are the reaction time hazard distributions?

      These are excellent suggestions. We are looking into implementing them.

      It is not clear that the animals on this task were actually using cognitive control strategies on this task. One cannot assume from the task that cognitive control is key. The authors only consider a very limited number of potential behaviors (an overly simple RL model). On this task, there are a lot of potential behavioral strategies: "win-stay/lose-shift", "perseveration", "alternation", even "random choices" should be considered.

      The strategies the Reviewer mentioned are descriptors of the actual choices the rats made. For example, perseveration means the rat is choosing one of the levers at an excessively high rate whereas alternation means it is choosing the two levers more or less equally, independent of payouts. But the question we are interested in is why? We are arguing that the type of cognitive control determines the choice behavior but cognitive control is an internal variable that guides behavior, rather than simply a descriptor of the behavior. For example, the animal opts to perseverate on the delayed lever because the cognitive control required to track ival is too high. We then searched the neural data for signatures of the two types of cognitive control.

      The delay lever was assigned to the "non-preferred side". How did side bias affect the decisions made?

      The side bias clearly does not impact performance as the animals prefer the delay lever at shorter delays, which works against this bias.

      The analyses based on "group" are unjustified. The authors compare the proportion of delayed to immediate lever press choices on the non-forced trials and then did k-means clustering on this distribution. But the distribution itself was not shown, so it is unclear whether the "groups" were actually different. They used k=3, but do not describe how this arbitrary number was chosen. (Is 3 the optimal number of clusters to describe this distribution?) Moreover, they removed three group 1 sessions with an 8s delay and two group 2 sessions with a 4s delay, making all the group 1 sessions 4s delay sessions and all group 2 sessions 8s delay sessions. They then ignore group 3 completely. These analyses seem arbitrary and unnecessarily complex. I think they need to analyze the data by delay. (How do rats handle 4s delay sessions? How do rats handle 6s delay sessions? How do rats handle 8s delay sessions?). If they decide to analyze the data by strategy, then they should identify specific strategies, model those strategies, and do model comparison to identify the best explanatory strategy. Importantly, the groups were session-based, not rat based, suggesting that rats used different strategies based on the delay to the delayed lever.

      These are excellent points and, as stated above, we are in the process revisiting the group assignments in an effort allay these criticisms.

      The reinforcement learning model used was overly simple. In particular, the RL model assumes that the subjects understand the task structure, but we know that even humans have trouble following complex task structures. Moreover, we know that rodent decision-making depends on much more complex strategies (model-based decisions, multi-state decisions, rate-based decisions, etc). There are lots of other ways to encode these decision variables, such as softmax with an inverse temperature rather than epsilon-greedy. The RL model was stated as a given and not justified. As one critical example, the RL model fit to the data assumed a constant exponential discounting function, but it is well-established that all animals, including rodents, use hyperbolic discounting in intertemporal choice tasks. Presumably this changes dramatically the effect of 4s and 8s. As evidence that the RL model is incomplete, the parameters found for the two groups were extreme. (Alpha=1 implies no history and only reacting to the most recent event. Epsilon=0.4 in an epsilon-greedy algorithm is a 40% chance of responding randomly.)

      Please see our response above. We agree that the approach was not justified, but we do not agree that it is invalid. Simply stated, a softmax approach gives the best fit to the choice behavior, whereas our epsilon-greedy approach attempted to reproduce the choice behavior using a naïve agent that progressively learns the values of the two levers on a choice-by-choice basis. The epsilon-greedy approach can therefore tell us whether it is possible to reproduce the choice behavior by an agent that is only tracking ival. Given our discovery of an ival-tracking signal in ACC, we believed that this was a critical point (although admittedly we did a poor job of communicating it). However, we also appreciate that important insights can be gained by fitting a model to the data as suggested. In fact, we had implemented this approach initially and are currently reconsidering what it can tell us in light of the Reviewers comments.

      The authors do add a "dbias" (which is a preference for the delayed lever) term to the RL model, but note that it has to be maximal in the 4s condition to reproduce group 2 behavior, which means they are not doing reinforcement learning anymore, just choosing the delayed lever.

      Exactly. The model results indicated that a naïve agent that relied only on ival tracking would not behave in this manner. Hence it therefore was unlikely that the G1 animals were using an ival-tracking strategy, even though a strong ival-tracking signal was present in ACC.

      Neurophysiology:

      The neurophysiology figures are unclear and mostly uninterpretable; they do not show variability, statistics or conclusive results.

      While the reviewer is justified in criticizing the clarity of the figures, the statement that “they do not show variability, statistics or conclusive results” is demonstrably false. Each of the figures presented in the manuscript, except Figure 3, are accompanied by statistics and measures of variability. This comment is hyperbolic and not justified.  

      Figure 3 was an attempt to show raw neural data to better demonstrate how robust the ivalue tracking signal is.

      As with the behavior, I would have liked to have seen more traditional neurophysiological analyses first. What do the cells respond to? How do the manifolds change aligned to the lever presses? Are those different between lever presses?

      We provide several figures describing how neurons change firing rates in response to varying reward. We are unsure what the reviewer means by “traditional analysis”, especially since this is immediately followed by a request for an assessment of neural manifolds. That said, we are developing ways to make the analysis more intuitive and, hopefully, more “traditional”.

      Are there changes in cellular information (both at the individual and ensemble level) over time in the session?

      We provide several analyses of how firing rate changes over trials in relation to ival over time in the session.

      How do cellular responses differ during that delay while both levers are out, but the rats are not choosing the immediate lever?

      It is not clear to us how this analysis addresses our hypothesis regarding control signals in ACC.

      Figure 3, for example, claims that some of the principal components tracked the number of pellets on the immediate lever ("ival"), but they are just two curves. No statistics, controls, or justification for this is shown. BTW, on Figure 3, what is the event at 200s?

      Figure 3 will be folded into one of the other figures that contains the summary statistics.

      I'm confused. On Figure 4, the number of trials seems to go up to 50, but in the methods, they say that rats received 40 trials or 45 minutes of experience.

      This analysis included force trials. The max of the session is 40 choice trials. We will clarify in the revised manuscript. 

      At the end of page 14, the authors state that the strength of the correlation did not differ by group and that this was "predicted" by the RL modeling, but this statement is nonsensical, given that the RL modeling did not fit the data well, depended on extreme values. Moreover, this claim is dependent on "not statistically detectable", which is, of course, not interpretable as "not different".

      We plan to revisit this analysis and the RL model.

      There is an interesting result on page 16 that the increases in theta power were observed before a delayed lever press but not an immediate lever press, and then that the theta power declined after an immediate lever press.

      Thank you for the positive comment.

      These data are separated by session group (again group 1 is a subset of the 4s sessions, group 2 is a subset of the 8s sessions, and group 3 is ignored). I would much rather see these data analyzed by delay itself or by some sort of strategy fit across delays.

      Provisional analysis indicates that the results hold up over delays, rather than the groupings in the paper. We will address this in a full revision of the manuscript.

      That being said, I don't see how this description shows up in Figure 6. What does Figure 6 look like if you just separate the sessions by delay?

      We are unclear what the reviewer means by “this description”.

      Discussion:

      Finally, it is unclear to what extent this task actually gets at the questions originally laid out in the goals and returned to in the discussion. The idea of cognitive effort is interesting, but there is no data presented that this task is cognitive at all. The idea of a resourced cognitive effort and a resistance cognitive effort is interesting, but presumably the way one overcomes resistance is through resource-limited components, so it is unclear that these two cognitive effort strategies are different.

      We view the strong evidence for ival tracking presented herein as a potentially critical component of resource based cognitive effort. We hope to clarify how this task engaged cognitive effort more clearly.  

      The authors state that "ival-tracking" (neurons and ensembles that presumably track the number of pellets being delivered on the immediate lever - a fancy name for "expectations") "taps into a resourced-based form of cognitive effort", but no evidence is actually provided that keeping track of the expectation of reward on the immediate lever depends on attention or mnemonic resources. They also state that a "dLP-biased strategy" (waiting out the delay) is a "resistance-based form of cognitive effort" but no evidence is made that going to the delayed side takes effort.

      There is a well-developed literature that rats and mice do not like waiting for delayed reinforcers. We contend that enduring something you don’t like takes effort.

      The authors talk about theta synchrony, but never actually measure theta synchrony, particularly across structures such as amygdala or ventral hippocampus. The authors try to connect this to "the unpleasantness of the delay", but provide no measures of pleasantness or unpleasantness. They have no evidence that waiting out an 8s delay is unpleasant.

      We will better clarify how our measure of Theta power relates to synchrony. There is a well-developed literature that rats and mice do not like waiting for delayed reinforcers.

      The authors hypothesize that the "ival-tracking signal" (the expectation of number of pellets on the immediate lever) "could simply reflect the emotional or autonomic response". Aside from the fact that no evidence for this is provided, if this were to be true, then, in what sense would any of these signals be related to cognitive control?

      This is proposed as an alternative explanation to the ivalue signal. We provide this as a possibility, never a conclusion. We will clarify this in the revised text. 

      Reviewer #2 (Public Review):

      Summary:

      This manuscript explores the neuronal signals that underlie resistance vs resource-based models of cognitive effort. The authors use a delayed discounting task and computational models to explore these ideas. The authors find that the ACC strongly tracks value and time, which is consistent with prior work. Novel contributions include quantification of a resource-based control signal among ACC ensembles, and linking ACC theta oscillations to a resistance-based strategy.

      Strengths:

      The experiments and analyses are well done and have the potential to generate an elegant explanatory framework for ACC neuronal activity. The inclusion of local-field potential / spike-field analyses is particularly important because these can be measured in humans.

      Thank you for the endorsement of our work.

      Weaknesses:

      I had questions that might help me understand the task and details of neuronal analyses.

      (1) The abstract, discussion, and introduction set up an opposition between resource and resistance based forms of cognitive effort. It's clear that the authors find evidence for each (ACC ensembles = resource, theta=resistance?) but I'm not sure where the data fall on this dichotomy.

      a. An overall very simple schematic early in the paper (prior to the MCML model? or even the behavior) may help illustrate the main point.

      b. In the intro, results, and discussion, it may help to relate each point to this dichotomy.

      c. What would resource-based signals look like? What would resistance based signals look like? Is the main point that resistance-based strategies dominate when delays are short, but resource-based strategies dominate when delays are long?

      d. I wonder if these strategies can be illustrated? Could these two measures (dLP vs ival tracking) be plotted on separate axes or extremes, and behavior, neuronal data, LFP, and spectral relationships be shown on these axes? I think Figure 2 is working towards this. Could these be shown for each delay length? This way, as the evidence from behavior, model, single neurons, ensembles, and theta is presented, it can be related to this framework, and the reader can organize the findings.

      These are excellent suggestions, and we intend to implement each of them, where possible.

      (2) The task is not clear to me.

      a. I wonder if a task schematic and a flow chart of training would help readers.

      Yes, excellent idea, we intend to include this.

      b. This task appears to be relatively new. Has it been used before in rats (Oberlin and Grahame is a mouse study)? Some history / context might help orient readers.

      Indeed, this task has been used in rats in several prior studies in rats. Please see the following references (PMID: 39119916, 31654652, 28000083, 26779747, 12270518, 19389183).

      c. How many total sessions were completed with ascending delays? Was there criteria for surgeries? How many total recording sessions per animal (of the 54?)

      Please note that the delay does not change within a session. There was no criteria for surgery. In addition, we will update Table 1 to make the number of recording sessions more clear.

      d. How many trials completed per session (40 trials OR 45 minutes)? Where are there errors? These details are important for interpreting Figure 1.

      Every animal in this data set completed 40 trials. We will update the task description to clarify this issue. There are no errors in this task, but rather the task is designed to the tendency to make an impulsive choice (smaller reward now). We will provide clarity to this issue in the revision of the manuscript.   

      (3) Figure 1 is unclear to me.

      a. Delayed vs immediate lever presses are being plotted - but I am not sure what is red, and what is blue. I might suggest plotting each animal.

      We will clarify the colors and look into schemes to graph the data set.

      b. How many animals and sessions go into each data point?

      This information is in Table 1, but this could be clearer, and we will update the manuscript.

      c. Table 1 (which might be better referenced in the paper) refers to rats by session. Is it true that some rats (2 and 8) were not analyzed for the bulk of the paper? Some rats appear to switch strategies, and some stay in one strategy. How many neurons come from each rat?

      Table 1 is accurate, and we can add the number of neurons from each animal.

      d. Task basics - RT, choice, accuracy, video stills - might help readers understand what is going into these plots

      e. Does the animal move differently (i.e., RTs) in G1 vs. G2?

      We will look into ways to incorporate this information.

      (4) I wasn't sure how clustered G1 vs. G2 vs G3 are. To make this argument, the raw data (or some axis of it) might help.

      a. This is particularly important because G3 appears to be a mix of G1 and G2, although upon inspection, I'm not sure how different they really are

      b. Was there some objective clustering criteria that defined the clusters?

      c. Why discuss G3 at all? Can these sessions be removed from analysis?

      These are all excellent suggestions and points. We plan to revisit the strategy to assign sessions to groups, which we hope will address each of these points.

      (5) The same applies to neuronal analyses in Fig 3 and 4

      a. What does a single neuron peri-event raster look like? I would include several of these.

      b. What does PC1, 2 and 3 look like for G1, G2, and G3?

      c. Certain PCs are selected, but I'm not sure how they were selected - was there a criteria used? How was the correlation between PCA and ival selected? What about PCs that don't correlate with ival?

      d. If the authors are using PCA, then scree plots and PETHs might be useful, as well as comparisons to PCs from time-shuffled / randomized data.

      We will make several updates to enhance clarity of the neural data analysis, including adding more representative examples. We feel the need to balance the inclusion of representative examples with groups stats given the concerns raised by R1.

      (6) I had questions about the spectral analysis

      a. Theta has many definitions - why did the authors use 6-12 Hz? Does it come from the hippocampal literature, and is this the best definition of theta?. What about other bands (delta - 1-4 Hz), theta (4-7 Hz); and beta - 13- 30 Hz? These bands are of particular importance because they have been associated with errors, dopamine, and are abnormal in schizophrenia and Parkinson's disease.

      This designation comes mainly from the hippocampal and ACC literature in rodents. In addition, this range best captured the peak in the power spectrum in our data. Note that we focus our analysis on theta give the literature regarding theta in the ACC as a correlate of cognitive controls (references in manuscript). We did interrogate other bands as a sanity check and the results were mostly limited to theta. Given the scope of our manuscript and the concerns raised regarding complexity we are concerned that adding frequency analyses beyond theta obfuscates the take home message. However, we think this is worthy, and we will determine if this can be done in a brief, clear, and effective manner.

      b. Power spectra and time-frequency analyses may justify the authors focus. I would show these (y-axis - frequency, x-axis - time, z-axis, power).

      This is an excellent suggestion that we look forward to incorporating. 

      (7) PC3 as an autocorrelation doesn't seem the to be right way to infer theta entrainment or spike-field relationships, as PCA can be vulnerable to phantom oscillations, and coherence can be transient. It is also difficult to compare to traditional measures of phase-locking. Why not simply use spike-field coherence? This is particularly important with reference to the human literature, which the authors invoke.

      Excellent suggestion. We will look into the phantom oscillation issue. Note that PCA provided a way to classify neurons that exhibited peaks in the autocorrelation at theta frequencies. While spike-field coherence is a rigorous tool, it addresses a slightly different question (LFP entrainment). Notwithstanding, we plan to address this issue.  

      Reviewer #3 (Public Review):

      Summary:

      The study investigated decision making in rats choosing between small immediate rewards and larger delayed rewards, in a task design where the size of the immediate rewards decreased when this option was chosen and increased when it was not chosen. The authors conceptualise this task as involving two different types of cognitive effort; 'resistance-based' effort putatively needed to resist the smaller immediate reward, and 'resource-based' effort needed to track the changing value of the immediate reward option. They argue based on analyses of the behaviour, and computational modelling, that rats use different strategies in different sessions, with one strategy in which they consistently choose the delayed reward option irrespective of the current immediate reward size, and another strategy in which they preferentially choose the immediate reward option when the immediate reward size is large, and the delayed reward option when the immediate reward size is small. The authors recorded neural activity in anterior cingulate cortex (ACC) and argue that ACC neurons track the value of the immediate reward option irrespective of the strategy the rats are using. They further argue that the strategy the rats are using modulates their estimated value of the immediate reward option, and that oscillatory activity in the 6-12Hz theta band occurs when subjects use the 'resistance-based' strategy of choosing the delayed option irrespective of the current value of the immediate reward option. If solid, these findings will be of interest to researchers working on cognitive control and ACCs involvement in decision making. However, there are some issues with the experiment design, reporting, modelling and analysis which currently preclude high confidence in the validity of the conclusions.

      Strengths:

      The behavioural task used is interesting and the recording methods should enable the collection of good quality single unit and LFP electrophysiology data. The authors recorded from a sizable sample of subjects for this type of study. The approach of splitting the data into sessions where subjects used different strategies and then examining the neural correlates of each is in principle interesting, though I have some reservations about the strength of evidence for the existence of multiple strategies.

      Thank you for the positive comments.

      Weaknesses:

      The dataset is very unbalanced in terms of both the number of sessions contributed by each subject, and their distribution across the different putative behavioural strategies (see table 1), with some subjects contributing 9 or 10 sessions and others only one session, and it is not clear from the text why this is the case. Further, only 3 subjects contribute any sessions to one of the behavioural strategies, while 7 contribute data to the other such that apparent differences in brain activity between the two strategies could in fact reflect differences between subjects, which could arise due to e.g. differences in electrode placement. To firm up the conclusion that neural activity is different in sessions where different strategies are thought to be employed, it would be important to account for potential cross-subject variation in the data. The current statistical methods don't do this as they all assume fixed effects (e.g. using trials or neurons as the experimental unit and ignoring which subject the neuron/trial came from).

      This is an important issue that we plan to address with additional analysis in the manuscript update.

      It is not obvious that the differences in behaviour between the sessions characterised as using the 'G1' and 'G2' strategies actually imply the use of different strategies, because the behavioural task was different in these sessions, with a shorter wait (4 seconds vs 8 seconds) for the delayed reward in the G1 strategy sessions where the subjects consistently preferred the delayed reward irrespective of the current immediate reward size. Therefore the differences in behaviour could be driven by difference in the task (i.e. external world) rather than a difference in strategy (internal to the subject). It seems plausible that the higher value of the delayed reward option when the delay is shorter could account for the high probability of choosing this option irrespective of the current value of the immediate reward option, without appealing to the subjects using a different strategy.

      Further, even if the differences in behaviour do reflect different behavioural strategies, it is not obvious that these correspond to allocation of different types of cognitive effort. For example, subjects' failure to modify their choice probabilities to track the changing value of the immediate reward option might be due simply to valuing the delayed reward option higher, rather than not allocating cognitive effort to tracking immediate option value (indeed this is suggested by the neural data). Conversely, if the rats assign higher value to the delayed reward option in the G1 sessions, it is not obvious that choosing it requires overcoming 'resistance' through cognitive effort.

      The RL modelling used to characterise the subject's behavioural strategies made some unusual and arguably implausible assumptions:

      i) The goal of the agent was to maximise the value of the immediate reward option (ival), rather than the standard assumption in RL modelling that the goal is to maximise long-run (e.g. temporally discounted) reward. It is not obvious why the rats should be expected to care about maximising the value of only one of their two choice options rather than distributing their choices to try and maximise long run reward.

      ii) The modelling assumed that the subject's choice could occur in 7 different states, defined by the history of their recent choices, such that every successive choice was made in a different state from the previous choice. This is a highly unusual assumption (most modelling of 2AFC tasks assumes all choices occur in the same state), as it causes learning on one trial not to generalise to the next trial, but only to other future trials where the recent choice history is the same.

      iii) The value update was non-standard in that rather than using the trial outcome (i.e. the amount of reward obtained) as the update target, it instead appeared to use some function of the value of the immediate reward option (it was not clear to me from the methods exactly how the fival and fqmax terms in the equation are calculated) irrespective of whether the immediate reward option was actually chosen.

      iv) The model used an e-greedy decision rule such that the probability of choosing the highest value option did not depend on the magnitude of the value difference between the two options. Typically, behavioural modelling uses a softmax decision rule to capture a graded relationship between choice probability and value difference.

      v) Unlike typical RL modelling where the learned value differences drive changes in subjects' choice preferences from trial to trial, to capture sensitivity to the value of the immediately rewarding option the authors had to add in a bias term which depended directly on this value (not mediated by any trial-to-trial learning). It is not clear how the rat is supposed to know the current trial ival if not by learning over previous trials, nor what purpose the learning component of the model serves if not to track the value of the immediate reward option.

      Given the task design, a more standard modelling approach would be to treat each choice as occurring in the same state, with the (temporally discounted) value of the outcomes obtained on each trial updating the value of the chosen option, and choice probabilities driven in a graded way (e.g. softmax) by the estimated value difference between the options. It would be useful to explicitly perform model comparison (e.g. using cross-validated log-likelihood with fitted parameters) of the authors proposed model against more standard modelling approaches to test whether their assumptions are justified. It would also be useful to use logistic regression to evaluate how the history of choices and outcomes on recent trials affects the current trial choice, and compare these granular aspects of the choice data with simulated data from the model.

      Each of the issues outlined above with the RL model a very important. We are currently re-evaluating the RL modeling approach in light of these comments. Please see comments to R1 regarding the model as they are relevant for this as well.

      There were also some issues with the analyses of neural data which preclude strong confidence in their conclusions:

      Figure 4I makes the striking claim that ACC neurons track the value of the immediately rewarding option equally accurately in sessions where two putative behavioural strategies were used, despite the behaviour being insensitive to this variable in the G1 strategy sessions. The analysis quantifies the strength of correlation between a component of the activity extracted using a decoding analysis and the value of the immediate reward option. However, as far as I could see this analysis was not done in a cross-validated manner (i.e. evaluating the correlation strength on test data that was not used for either training the MCML model or selecting which component to use for the correlation). As such, the chance level correlation will certainly be greater than 0, and it is not clear whether the observed correlations are greater than expected by chance.

      This is an astute observation and we plan to address this concern. We agree that cross-validation may provide an appropriate tool here.

      An additional caveat with the claim that ACC is tracking the value of the immediate reward option is that this value likely correlates with other behavioural variables, notably the current choice and recent choice history, that may be encoded in ACC. Encoding analyses (e.g. using linear regression to predict neural activity from behavioural variables) could allow quantification of the variance in ACC activity uniquely explained by option values after controlling for possible influence of other variables such as choice history (e.g. using a coefficient of partial determination).

      This is also an excellent point that we plan to address the manuscript update.

      Figure 5 argues that there are systematic differences in how ACC neurons represent the value of the immediate option (ival) in the G1 and G2 strategy sessions. This is interesting if true, but it appears possible that the effect is an artefact of the different distribution of option values between the two session types. Specifically, due to the way that ival is updated based on the subjects' choices, in G1 sessions where the subjects are mostly choosing the delayed option, ival will on average be higher than in G2 sessions where they are choosing the immediate option more often. The relative number of high, medium and low ival trials in the G1 and G2 sessions will therefore be different, which could drive systematic differences in the regression fit in the absence of real differences in the activity-value relationship. I have created an ipython notebook illustrating this, available at: https://notebooksharing.space/view/a3c4504aebe7ad3f075aafaabaf93102f2a28f8c189ab9176d4807cf1565f4e3. To verify that this is not driving the effect it would be important to balance the number of trials at each ival level across sessions (e.g. by subsampling trials) before running the regression.

      Excellent point and thank you for the notebook. We explored a similar approach previously but did not pursue it to completion. We will re-investigate this issue.

    1. Welcome to this lesson, where I'm going to very briefly talk about a special type of IAM role, and that's service-linked roles.

      Now, luckily there isn't a great deal of difference between service-linked roles and IAM roles. They're just used in a very specific set of situations. So let's jump in and get started.

      So simply put, a service-linked role is an IAM role linked to a specific AWS service. They provide a set of permissions which is predefined by a service. These permissions allow a single AWS service to interact with other AWS services on your behalf.

      Now, service-linked roles might be created by the service itself, or the service might allow you to create the role during the setup process of that service. Service-linked roles might also get created within IAM.

      The key difference between service-linked roles and normal roles is that you can't delete a service-linked role until it's no longer required. This means it must no longer be used within that AWS service. So that's the one key difference.

      In terms of permissions needed to create a service-linked role, here's an example of a policy that allows you to create a service-linked role.

      You'll notice a few key elements in this policy. The top statement is an allow statement. The action is iam:CreateServiceLinkedRole. For the resource, it has SERVICE-NAME.amazonaws.com.

      The important thing here is not to try to guess this, as different services express this in different ways. The formatting can differ, and it's case-sensitive. I've included a link with an overview of these details attached to this lesson.

      When creating this type of policy to allow someone to create service-linked roles, you have to be careful to ensure you do not guess this element of a statement.

      Another important consideration with service-linked roles is role separation. When I talk about role separation, I'm not using it in a technical sense, but in a job role sense.

      Role separation is where you might give one group of people the ability to create roles and another group the ability to use them. For instance, we might want to give Bob, one of our users, the ability to use a service-linked role with an AWS service.

      This involves using the architecture of being able to take a service-linked role and assign it to a service. If you want to give Bob the ability to use a preexisting role with a service but not create or edit that role, you would need to provide Bob with PassRole permissions. This allows Bob to pass an existing role into an AWS service. It's an example of role separation, meaning Bob could configure a service with a role that has already been created by a member of the security team. Bob would just need ListRole and PassRole permissions on that specific role.

      This is similar to when you use a pre-created role, for example, with a CloudFormation stack. By default, when creating a CloudFormation stack, CloudFormation uses the permissions of your identity to interact with AWS. This means you need permissions not only to create a stack but also to create the resources that the stack creates. However, you can give, for example, a user like Bob the ability to pass a role into CloudFormation. That role could have permissions that exceed those which Bob directly has. So a role that Bob uses could have the ability to create AWS resources that Bob does not. Bob might have access to create a stack and pass in a role, but the role provides CloudFormation with the permissions needed to interact with AWS.

      PassRole is a method inside AWS that allows you to implement role separation, and it's something you can also use with service-linked roles. This is something I wanted to reiterate to emphasize that passing a role is a very important AWS security architecture.

      That is everything I wanted to cover in this very brief lesson. It's really just an extension of what you've already learned about IAM roles, and it's something you'll use in demo lessons elsewhere in the course.

      For now, I just want you to be aware of how service-linked roles differ from normal roles and how the PassRole architecture works. With that being said, that's everything I wanted to cover in this video.

      So go ahead and complete the video, and when you're ready, I look forward to you joining me in the next.

    1. “cruel optimism”

      I think this is a very powerful concept -- as a former student said, entrepreneurs must have cruel optimism, individuals must be able to adapt and ride out whatever wave or obstacle comes around.

      Berlant uses the term cruel optimism to refer to our our investments in “compromised conditions of possibility whose realization is discovered to be impossible, sheer fantasy.” (i.e., we keep cheering for a team we know will lose; we maintain hope in an unattainable romantic ideal promulgated by Hollywood or pursue happiness based on unrealistic beauty standards; we engage in small acts of environmental stewardship like recycling or buying a hybrid in the face of potentially unstoppable climate change...) Berlant basically means that the thing we seek to achieve, the thing (or state of being) that we desire (or the act of seeking and desiring itself) might actually threaten our well-being (that's what makes it cruel!). As she put it succinctly, “a relation of cruel optimism exists when something you desire is actually an obstacle to your flourishing.”

      This relates to entrepreneurialism in so many ways: Engaging in the gig economy or a side-hustle as a way to increase one's income (or security) in uncertain times is cruel and optimistic. Similarly, we encounter aspirational labour in the form of internships or any form of unpaid labour while looking for a "real" job. Perhaps you feel the pressure of cultivating a sense of employability. According to Frayne (2015), today, students are expected “to improve their prospects by training, acquiring educational credentials, networking, learning how to project the right kind of personality, and gaining life experiences that match up with the values sought by employers.” In other words, they have to act entrepreneurially even to get a non-entrepreneurial job. As Hawzen et al. (2018) assert, this incites anxiety and results in a colonization of one’s entire life by work-related demands as students feel the need to separate themselves from the competition, doing things like volunteering to gain an advantage or to get a "foot in the door"... We also see it to a certain extent in the example of entrepreneurial vloggers in the sense that the fantasy of a “good life” through fame and fortune is rarely realized. The cruel conditions of precocity are, for most, more of a reality than the fantasy... and we take up this theme explicitly in two weeks hence with digital 'autopreneurs'

      Overall, this also highlights one of the reasons I chose this article -- rather than just highlighting how entrepreneurs are certain types of people (or motivated by certain types of things), it emphasizes how entrepreneurship is a mental orientation, not just a business concept but a way of living. But it's not all sunshine and happiness. Cruel optimism, indeed!

      What about you? Are you familiar with the feeling of 'cruel optimism'? Does it define the current times or your current disposition?

    2. The status of entrepreneurship as a new common sense of subjectivity and economic practice

      Remember at the beginning of the article (when Szeman says "we are all entrepreneurs now") (p. 472)? He doesn't mean that we are all creating business start-ups. Rather, he's suggesting that there is a spirit-of-the-times wherein entrepreneurship has become this new common-sense reality. It is both a dominant way of thinking about how we ought to act, AND an informal rulebook for how economies (and other forms of practice) ought to function too... In other words, entrepreneurship isn't just about undertaking profit-making (and risk-inducing) economic practices in capitalism. Rather, it's about undertaking a new subjectivity, a new identity when it comes to how we think of ourselves, how we relate to others, and how we respond to our wider social, cultural, political, and economic environment.

    1. Welcome back and welcome to this CloudTrail demo where we're going to set up an organizational trail and configure it to log data for all accounts in our organization to S3 and CloudWatch logs.

      The first step is that you'll need to be logged into the IAM admin user of the management account of the organization. As a reminder, this is the general account. To set up an organizational trail, you always need to be logged into the management account. To set up individual trails, you can do that locally inside each of your accounts, but it's always more efficient to use an organizational trail.

      Now, before we start the demonstration, I want to talk briefly about CloudTrail pricing. I'll make sure this link is in the lesson description, but essentially there is a fairly simple pricing structure to CloudTrail that you need to be aware of.

      The 90-day history that's enabled by default in every AWS account is free. You don't get charged for that; it comes free by default with every AWS account. Next, you have the ability to get one copy of management events free in every region in each AWS account. This means creating one trail that's configured for management events in each region in each AWS account, and that comes for free. If you create any additional trails, so you get any additional copies of management events, they are charged at two dollars per 100,000 events. That won't apply to us in this demonstration, but you need to be aware of that if you're using this in production.

      Logging data events comes at a charge regardless of the number, so we're not going to enable data events for this demo lesson. But if you do enable it, then that comes at a charge of 10 cents per 100,000 events, irrespective of how many trails you have. This charge applies from the first time you're logging any data events.

      What we'll be doing in this demo lesson is setting up an organizational trail which will create a trail in every region in every account inside the organization. But because we get one for free in every region in every account, we won't incur any charges for the CloudTrail side of things. We will be charged for any S3 storage that we use. However, S3 also comes with a free tier allocation for storage, which I don't expect us to breach.

      With that being said, let's get started and implement this solution. To do that, we need to be logged in to the console UI again in the management account of the organization. Then we need to move to the CloudTrail console. If you've been here recently, it will be in the Recently Visited Services. If not, just type CloudTrail in the Find Services box and then open the CloudTrail console.

      Once you're at the console, you might see a screen like this. If you do, then you can just click on the hamburger menu on the left and then go ahead and click on trails. Now, depending on when you're doing this demo, if you see any warnings about a new or old console version, make sure that you select the new version so your console looks like what's on screen now.

      Once you're here, we need to create a trail, so go ahead and click on create trail. To create a trail, you're going to be asked for a few important pieces of information, the first of which is the trail name. For trail name, we're going to use "animals4life.org," so just go ahead and enter that. By default, with this new UI version, when you create a trail, it's going to create it in all AWS regions in your account. If you're logged into the management account of the organization, as we are, you also have the ability to enable it for all regions in all accounts of your organization. We're going to do that because this allows us to have one single logging location for all CloudTrail logs in all regions in all of our accounts, so go ahead and check this box.

      By default, CloudTrail stores all of its logs in an S3 bucket. When you're creating a trail, you have the ability to either create a new S3 bucket to use or you can use an existing bucket. We're going to go ahead and create a brand new bucket for this trail. Bucket names within S3 need to be globally unique, so it needs to be a unique name across all regions and across all AWS accounts. We're going to call this bucket starting with "CloudTrail," then a hyphen, then "animals-for-life," another hyphen, and then you'll need to put a random number. You’ll need to pick something different from me and different from every other student doing this demo. If you get an error about the bucket name being in use, you just need to change this random number.

      You're also able to specify if you want the log files stored in the S3 bucket to be encrypted. This is done using SSE-KMS encryption. This is something that we'll be covering elsewhere in the course, and for production usage, you would definitely want to use it. For this demonstration, to keep things simple, we're not going to encrypt the log files, so go ahead and untick this box.

      Under additional options, you're able to select log file validation, which adds an extra layer of security. This means that if any of the log files are tampered with, you have the ability to determine that. This is a really useful feature if you're performing any account-level audits. In most production situations, I do enable this, but you can also elect to have an SNS notification delivery. So, every time log files are delivered into this S3 bucket, you can have a notification. This is useful for production usage or if you need to integrate this with any non-AWS systems, but for this demonstration, we'll leave this one unchecked.

      You also have the ability, as well as storing these log files into S3, to store them in CloudWatch logs. This gives you extra functionality because it allows you to perform searches, look at the logs from a historical context inside the CloudWatch logs user interface, as well as define event-driven processes. You can configure CloudWatch logs to scan these CloudTrail logs and, in the event that any particular piece of text occurs in the logs (e.g., any API call, any actions by a user), you can generate an event that can invoke, for example, a Lambda function or spawn some other event-driven processing. Don't worry if you don't understand exactly what this means at this point; I'll be talking about all of this functionality in detail elsewhere in the course. For this demonstration, we are going to enable CloudTrail to put these logs into CloudWatch logs as well, so check this box. You can choose a log group name within CloudWatch logs for these CloudTrail logs. If you want to customize this, you can, but we're going to leave it as the default.

      As with everything inside AWS, if a service is acting on our behalf, we need to give it the permissions to interact with other AWS services, and CloudTrail is no exception. We need to give CloudTrail the ability to interact with CloudWatch logs, and we do that using an IAM role. Don’t worry, we’ll be talking about IAM roles in detail elsewhere in the course. For this demonstration, just go ahead and select "new" because we're going to create a new IAM role that will give CloudTrail the ability to enter data into CloudWatch logs.

      Now we need to provide a role name, so go ahead and enter "CloudTrail_role_for_CloudWatch_logs" and then an underscore and then "animals_for_life." The name doesn’t really matter, but in production settings, you'll want to make sure that you're able to determine what these roles are for, so we’ll use a standard naming format. If you expand the policy document, you'll be able to see the exact policy document or IAM policy document that will be used to give this role the permissions to interact with CloudWatch logs. Don’t worry if you don’t fully understand policy documents at this point; we’ll be using them throughout the course, and over time you'll become much more comfortable with exactly how they're used. At a high level, this policy document will be attached to this role, and this is what will give CloudTrail the ability to interact with CloudWatch logs.

      At this point, just scroll down; that's everything that we need to do, so go ahead and click on "next." Now, you'll need to select what type of events you want this trail to log. You’ve got three different choices. The default is to log only management events, so this logs any events against the account or AWS resources (e.g., starting or stopping an EC2 instance, creating or deleting an EBS volume). You've also got data events, which give you the ability to log any actions against things inside resources. Currently, CloudTrail supports a wide range of services for data event logging. For this demonstration, we won't be setting this up with data events initially because I’ll be covering this elsewhere in the course. So, go back to the top and uncheck data events.

      You also have the ability to log insight events, which can identify any unusual activity, errors, or user behavior on your account. This is especially useful from a security perspective. For this demonstration, we won’t be logging any insight events; we’re just going to log management events. For management events, you can further filter down to read or write or both and optionally exclude KMS or RDS data API events. For this demo lesson, we’re just going to leave it as default, so make sure that read and write are checked. Once you've done that, go ahead and click on "next." On this screen, just review everything. If it all looks good, click on "create trail."

      Now, if you get an error saying the S3 bucket already exists, you'll just need to choose a new bucket name. Click on "edit" at the top, change the bucket name to something that's globally unique, and then follow that process through again and create the trail.

      Certainly! Here is the continuation and completion of the transcript:


      After a few moments, the trail will be created. It should say "US East Northern Virginia" as the home region. Even though you didn't get the option to select it because it's selected by default, it is a multi-region trail. Finally, it is an organizational trail, which means that this trail is now logging any CloudTrail events from all regions in all accounts in this AWS organization.

      Now, this isn't real-time, and when you first enable it, it can take some time for anything to start to appear in either S3 or CloudWatch logs. At this stage, I recommend that you pause the video and wait for 10 to 15 minutes before continuing, because the initial delivery of that first set of log files through to S3 can take some time. So pause the video, wait 10 to 15 minutes, and then you can resume.

      Next, right-click the link under the S3 bucket and open that in a new tab. Go to that tab, and you should start to see a folder structure being created inside the S3 bucket. Let's move down through this folder structure, starting with CloudTrail. Go to US East 1 and continue down through this folder structure.

      In my case, I have quite a few of these log files that have been delivered already. I'm going to pick one of them, the most recent, and just click on Open. Depending on the browser that you're using, you might have to download and then uncompress this file. Because I'm using Firefox, it can natively open the GZ compressed file and then automatically open the JSON log file inside it.

      So this is an example of a CloudTrail event. We're able to see the user identity that actually generates this event. In this case, it's me, I am admin. We can see the account ID that this event is for. We can see the event source, the event name, the region, the source IP address, the user agent (in this case, the console), and all of the relevant information for this particular interaction with the AWS APIs are logged inside this CloudTrail event.

      Don’t worry if this doesn’t make a lot of sense at this point. You’ll get plenty of opportunities to interact with this type of logging event as you go through the various theory and practical lessons within the course. For now, I just want to highlight exactly what to expect with CloudTrail logs.

      Since we’ve enabled all of this logging information to also go into CloudWatch logs, we can take a look at that as well. So back at the CloudTrail console, if we click on Services and then type CloudWatch, wait for it to pop up, locate Logs underneath CloudWatch, and then open that in a new tab.

      Inside CloudWatch, on the left-hand menu, look for Logs, and then Log Groups, and open that. You might need to give this a short while to populate, but once it does, you should see a log group for the CloudTrail that you’ve just created. Go ahead and open that log group.

      Inside it, you’ll see a number of log streams. These log streams will start with your unique organizational code, which will be different for you. Then there will be the account number of the account that it represents. Again, these will be different for you. And then there’ll be the region name. Because I’m only interacting with the Northern Virginia region, currently, the only ones that I see are for US East 1.

      In this particular account that I’m in, the general account of the organization, if I look at the ARN (Amazon Resource Name) at the top or after US East 1 here, this number is my account number. This is the account number of my general account. So if I look at the log streams, you’ll be able to see that this account (the general account) matches this particular log stream. You’ll be able to do the same thing in your account. If you look for this account ID and then match it with one of the log streams, you'll be able to pull the logs for the general AWS account.

      If I go inside this particular log stream, as CloudTrail logs any activity in this account, all of that information will be populated into CloudWatch logs. And that’s what I can see here. If I expand one of these log entries, we’ll see the same formatted CloudTrail event that I just showed you in my text editor. So the only difference when using CloudWatch logs is that the CloudTrail events also get entered into a log stream in a log group within CloudWatch logs. The format looks very similar.

      Returning to the CloudTrail console, one last thing I want to highlight: if you expand the menu on the left, whether you enable a particular trail or not, you’ve always got access to the event history. The event history stores a log of all CloudTrail events for the last 90 days for this particular account, even if you don’t have a specific trail enabled. This is standard functionality. What a trail allows you to do is customize exactly what happens to that data. This area of the console, the event history, is always useful if you want to search for a particular event, maybe check who’s logged onto the account recently, or look at exactly what the IAM admin user has been doing within this particular AWS account.

      The reason why we created a trail is to persistently store that data in S3 as well as put it into CloudWatch logs, which gives us that extra functionality. With that being said, that’s everything I wanted to cover in this demo lesson.

      One thing you need to be aware of is that S3, as a service, provides a certain amount of resource under the free tier available in every new AWS account, so you can store a certain amount of data in S3 free of charge. The problem with CloudTrail, and especially organizational trails, is that they generate quite a large number of requests. There is also, in addition to space, a number of requests per month that are part of the free tier.

      If you leave this CloudTrail enabled for the duration of your studies, for the entire month, it is possible that this will go slightly over the free tier allocation for requests within the S3 service. You might see warnings that you’re approaching a billable threshold, and you might even get a couple of cents of bill per month if you leave this enabled all the time. To avoid that, if you just go to Trails, open up the trail that you’ve created, and then click on Stop Logging. You’ll need to confirm that by clicking on Stop Logging, and at that point, no logging will occur into the S3 bucket or into CloudWatch logs, and you won’t experience those charges.

      For any production usage, the low cost of this service means that you would normally leave it enabled in all situations. But to keep costs within the free tier for this course, you can, if required, just go ahead and stop the logging. If you don’t mind a few cents per month of S3 charges for CloudTrail, then by all means, go ahead and leave it enabled.

      With that being said, that’s everything I wanted to cover in this demo lesson. So go ahead, complete the lesson, and when you're ready, I look forward to you joining me in the next.

    1. Welcome to this lesson, where I'm going to introduce the theory and architecture of CloudWatch Logs.

      I've already covered the metrics side of CloudWatch earlier in the course, and I'm covering the logs part now because you'll be using it when we cover CloudTrail. In the CloudTrail demo, we'll be setting up CloudTrail and using CloudWatch Logs as a destination for those logs. So, you'll need to understand it, and we'll be covering the architecture in this lesson. Let's jump in and get started.

      CloudWatch Logs is a public service. The endpoint to which applications connect is hosted in the AWS public zone. This means you can use the product within AWS VPCs, from on-premises environments, and even other cloud platforms, assuming that you have network connectivity as well as AWS permissions.

      The CloudWatch Logs product allows you to store, monitor, and access logging data. Logging data, at a very basic level, consists of a piece of information, data, and a timestamp. The timestamp generally includes the year, month, day, hour, minute, second, and timezone. There can be more fields, but at a minimum, it's generally a timestamp and some data.

      CloudWatch Logs has built-in integrations with many AWS services, including EC2, VPC Flow Logs, Lambda, CloudTrail, Route 53, and many more. Any services that integrate with CloudWatch Logs can store data directly inside the product. Security for this is generally provided by using IAM roles or service roles.

      For anything outside AWS, such as logging custom application or OS logs on EC2, you can use the unified CloudWatch agent. I’ve mentioned this before and will be demoing it later in the EC2 section of the course. This is how anything outside of AWS products and services can log data into CloudWatch Logs. So, it’s either AWS service integrations or the unified CloudWatch agent. There is a third way, using development kits for AWS to implement logging into CloudWatch Logs directly into your application, but that tends to be covered in developer and DevOps AWS courses. For now, just remember either AWS service integrations or the unified CloudWatch agent.

      CloudWatch Logs are also capable of taking logging data and generating a metric from it, known as a metric filter. Imagine a situation where you have a Linux instance, and one of the operating system log files logs any failed connection attempts via SSH. If this logging information was injected into CloudWatch Logs, a metric filter can scan those logs constantly. Anytime it sees a mention of the failed SSH connection, it can increment a metric within CloudWatch. You can then have alarms based on that metric, and I’ll be demoing that very thing later in the course.

      Let’s look at the architecture visually because I'll be showing you how this works in practice in the CloudTrail demo, which will be coming up later in the section. Architecturally, CloudWatch Logs looks like this: It’s a regional service. So, for this example, let’s assume we’re talking about us-east-1.

      The starting point is our logging sources, which can include AWS products and services, mobile or server-based applications, external compute services (virtual or physical servers), databases, or even external APIs. These sources inject data into CloudWatch Logs as log events.

      Log events consist of a timestamp and a message block. CloudWatch Logs treats this message as a raw block of data. It can be anything you want, but there are ways the data can be interpreted, with fields and columns defined. Log events are stored inside log streams, which are essentially a sequence of log events from the same source.

      For example, if you had a log file stored on multiple EC2 instances that you wanted to inject into CloudWatch Logs, each log stream would represent the log file for one instance. So, you’d have one log stream for instance one and one log stream for instance two. Each log stream is an ordered set of log events for a specific source.

      We also have log groups, which are containers for multiple log streams of the same type of logging. Continuing the example, we would have one log group containing everything for that log file. Inside this log group would be different log streams, each representing one source. Each log stream is a collection of log events. Every time an item was added to the log file on a single EC2 instance, there would be one log event inside one log stream for that instance.

      A log group also stores configuration settings, such as retention settings and permissions. When we define these settings on a log group, they apply to all log streams within that log group. It’s also where metric filters are defined. These filters constantly review any log events for any log streams in that log group, looking for certain patterns, such as an application error code or a failed SSH login. When detected, these metric filters increment a metric, and metrics can have associated alarms. These alarms can notify administrators or integrate with AWS or external systems to take action.

      CloudWatch Logs is a powerful product. This is the high-level architecture, but don’t worry—you’ll get plenty of exposure to it throughout the course because many AWS products integrate with CloudWatch Logs and use it to store their logging data. We’ll be coming back to this product time and again as we progress through the course. CloudTrail uses CloudWatch Logs, Lambda uses CloudWatch Logs, and VPC Flow Logs use CloudWatch Logs. There are many examples of AWS products where we’ll be integrating them with CloudWatch Logs.

      I just wanted to introduce it at this early stage of the course. That’s everything I wanted to cover in this theory lesson. Thanks for watching. Go ahead, complete this video, and when you’re ready, join me in the next.

    1. Welcome back, and in this lesson, I'll be talking about service control policies, or SCPs. SCPs are a feature of AWS Organizations which can be used to restrict AWS accounts. They're an essential feature to understand if you are involved in the design and implementation of larger AWS platforms. We've got a lot to cover, so let's jump in and get started.

      At this point, this is what our AWS account setup looks like. We've created an organization for Animals4life, and inside it, we have the general account, which from now on I'll be referring to as the management account, and then two member accounts, so production, which we'll call prod, and development, which we'll be calling dev. All of these AWS accounts are within the root container of the organization. That's to say they aren't inside any organizational units. In the next demo lesson, we're going to be adding organizational units, one for production and one for development, and we'll be putting the member accounts inside their respective organizational units.

      Now, let's talk about service control policies. The concept of a service control policy is simple enough. It's a policy document, a JSON document, and these service control policies can be attached to the organization as a whole by attaching them to the root container, or they can be attached to one or more organizational units. Lastly, they can even be attached to individual AWS accounts. Service control policies inherit down the organization tree. This means if they're attached to the organization as a whole, so the root container of the organization, then they affect all of the accounts inside the organization. If they're attached to an organizational unit, then they impact all accounts directly inside that organizational unit, as well as all accounts within OUs inside that organizational unit. If you have nested organizational units, then by attaching them to one OU, they affect that OU and everything below it. If you attach service control policies to one or more accounts, then they just directly affect those accounts that they're attached to.

      Now, I mentioned in an earlier lesson that the management account of an organization is special. One of the reasons it's special is that even if the management account has service control policies attached, either directly via an organizational unit, or on the root container of the organization itself, the management account is never affected by service control policies. This can be both beneficial and it can be a limitation, but as a minimum, you need to be aware of it as a security practice. Because the management account can't be restricted using service control policies, I generally avoid using the management account for any AWS resources. It's the only AWS account within AWS Organizations which can't be restricted using service control policies. As a takeaway, just remember that the management account is special and it's unaffected by any service control policies, which are attached to that account either directly or indirectly.

      Now, service control policies are account permissions boundaries. What I mean by that is they limit what the AWS account can do, including the Account Root User within that account. I talked earlier in the course about how you can't restrict an Account Root User. And that is true. You can't directly restrict what the Account Root User of an AWS account can do. The Account Root User always has full permissions over that entire AWS account, but with a service control policy, you're actually restricting what the account itself can do, specifically any identities within that account. So you're indirectly restricting the Account Root User because you're reducing the allowed permissions on the account; you're also reducing what the effective permissions on the Account Root User are. This is a really fine detail to understand. You can never restrict the Account Root User. It will always have 100% access to the account, but if you restrict the account, then in effect, you're also restricting the Account Root User.

      Now, you might apply a service control policy to prevent any usage of that account outside a known region, for example, us-east-1. You might also apply a service control policy which only allows a certain size of EC2 instance to be used within the account. Service control policies are a really powerful feature for any larger, more complex AWS deployments. The critical thing to understand about service control policies is they don't grant any permissions. Service control policies are just a boundary. They define the limit of what is and isn't allowed within the account, but they don't grant permissions. You still need to give identities within that AWS account permissions to AWS resources, but any SCPs will limit the permissions that can be assigned to individual identities.

      You can use service control policies in two ways. You can block by default and allow certain services, which is an allow list. Or you can allow by default and block access to certain services, which is a deny list. The default is a deny list. When you enable SCPs on your organization, AWS applies a default policy called full AWS access. This is applied to the organization and all OUs within that organization. This policy means that in the default implementation, service control policies have no effect since nothing is restricted. As a reminder, service control policies don't grant permissions, but when SCPs are enabled, there is an implicit default deny, just like IAM policies. If you had no initial allow, then everything would be denied. So the default is this full access policy, which essentially means no restrictions. It has the effect of making SCPs a deny list architecture, so you need to add any restrictions that you want to any AWS accounts within the organization. An example is that you could add another policy, such as this one, called DenyS3. This adds a deny policy for the entire S3 set of API operations, effectively denying S3. You need to remember that SCPs don't actually grant any access rights, but they establish which permissions can be granted in an account. The same priority rules apply: deny, allow, deny. Anything explicitly allowed in an SCP is a service which can have access granted to identities within that account, unless there's an explicit deny within an SCP, then a service cannot be granted. Explicit deny always wins. And in the absence of either, if we didn't have this full AWS access policy in place, then there would be an implicit deny, which blocks access to everything.

      The benefit of using deny lists is that because your foundation is to allow wildcard access, so all actions on all resources, as AWS extends the amounts of products and services which are available inside the platform, this allow list constantly expands to cover those services, so it's fairly low admin overhead. You simply need to add any services which you want to deny access to via an explicit deny. In certain situations, you might need to be more conscious about usage in your accounts, and that's where you'd use allow lists. To implement allow lists, it's a two-part architecture. One part of it is to remove the AWS full access policy. This means that only the implicit default deny is in place and active, and then you would need to add any services which you want to allow into a new policy. In this case, S3 and EC2. So in this architecture, we wouldn't have this full AWS access. We would be explicitly allowing S3 and EC2 access. So no matter what identity permissions identities in this account are provided with, they would only ever be allowed to access S3 and EC2. This is more secure because you have to explicitly say which services can be allowed access for users in those accounts, but it's much easier to make a mistake and block access to services which you didn't intend to. It's also much more admin overhead because you have to add services as your business requirements dictate. You can't simply have access to everything and deny services you don't want access to. With this type of architecture, you have to explicitly add each and every service which you want identities within the account to be able to access. Generally, I would normally suggest using a deny list architecture because, simply put, it's much lower admin overhead.

      Before we go into a demo, I want to visually show you how SCPs affect permissions. This is visually how SCPs impact permissions within an AWS account. In the left orange circle, this represents the different services that have been granted access to identities in an account using identity policies. On the right in red, this represents which services an SCP allows access to. So the SCP states that the three services in the middle and the service on the right are allowed access as far as the SCP is concerned, and the identity policies which were applied to identities within the account, so the orange circle on the left, grant access to four different services: the three in the middle and the one on the left.

      Only permissions which are allowed within identity policies in the account and are allowed by a service control policy are actually active. On the right, this access permission has no effect because while it's allowed within an SCP, an SCP doesn't grant access to anything; it just controls what can and can't be allowed by identity policies within that account. Because no identity policy allows access to this resource, then it has no effect. On the left, this particular access permission is allowed within an identity policy, but it's not effectively allowed because it's not allowed within an SCP. So only things which are involved, the identity policy and an SCP, are actually allowed. In this case, this particular access permission on the left has no effect because it's not within a service control policy, so it's denied.

      At an associate level, this is what you need to know for the exam. It's just simply understanding that your effective permissions for identities within an account are the overlap between any identity policies and any applicable SCPs. This is going to make more sense if you experience it with a demo, so this is what we're going to do next. Now that you've set up the AWS organization for the Animals4life business, it's time to put some of this into action. So I'm going to finish this lesson here and then in the next lesson, which is a demo, we're going to continue with the practical part of implementing SCPs. So go ahead and complete this video, and when you're ready, I'll look forward to you joining me in the next.

    1. Welcome to this lesson, where I'll be introducing AWS Organizations. AWS Organizations is a product that allows larger businesses to manage multiple AWS accounts in a cost-effective way with little to no management overhead.

      Organizations is a product that has evolved significantly over the past few years, and it's worthwhile to step through that evolution to understand all of its different features. We’ve got a lot to cover, so let's jump in and get started.

      Without AWS Organizations, many large businesses would face the challenge of managing numerous AWS accounts. In the example onscreen, there are four accounts, but I've worked with some larger enterprises with hundreds of accounts and have heard of even more. Without AWS Organizations, each of these accounts would have its own pool of IAM users as well as separate payment methods. Beyond 5 to 10 accounts, this setup becomes unwieldy very quickly.

      AWS Organizations is a simple product to understand. You start with a single AWS account, which I'll refer to as a standard AWS account from now on. A standard AWS account is an AWS account that is not part of an organization. Using this standard AWS account, you create an AWS Organization.

      It’s important to understand that the organization isn't created within this account; you're simply using the account to create the organization. This standard AWS account that you use to create the organization then becomes the Management Account for the organization. The Management Account used to be called the Master Account. If you hear either of these terms—Management Account or Master Account—just know that they mean the same thing.

      This is a key point to understand with regards to AWS Organizations because the Management Account is special for two reasons, which I’ll explain in this lesson. For now, I’ll add a crown to this account to indicate that it’s the Management Account and to help you distinguish it from other AWS accounts.

      Using this Management Account, you can invite other existing standard AWS accounts into the organization. Since these are existing accounts, they need to approve the invites to join the organization. Once they do, those Standard Accounts will become part of the AWS Organization.

      When standard AWS accounts join an AWS Organization, they change from being Standard Accounts to being Member Accounts of that organization. Organizations have one and only one Management or Master Account and then zero or more Member Accounts.

      You can create a structure of AWS accounts within an organization, which is useful if you have many accounts and need to group them by business units, functions, or even the development stage of an application. The structure within AWS Organizations is hierarchical, forming an inverted tree.

      At the top of this tree is the root container of the organization. This is just a container for AWS accounts at the top of the organizational structure. Don’t confuse this with the Account Root User, which is the admin user of an AWS account. The organizational root is just a container within an AWS Organization, which can contain AWS accounts, including Member Accounts or the Management Account.

      As well as containing accounts, the organizational root can also contain other containers, known as organizational units (OUs). These organizational units can contain AWS accounts, Member Accounts, or the Management Account, or they can contain other organizational units, allowing you to build a complex nested AWS account structure within Organizations.

      Again, please don’t confuse the organizational root with the AWS Account Root User. The AWS Account Root User is specific to each AWS account and provides full permissions over that account. The root of an AWS Organization is simply a container for AWS accounts and organizational units and is the top level of the hierarchical structure within AWS Organizations.

      One important feature of AWS Organizations is consolidated billing. With the example onscreen now, there are four AWS accounts, each with its own billing information. Once these accounts are added to an AWS Organization, the individual billing methods for the Member Accounts are removed. Instead, the Member Accounts pass their billing through to the Management Account of the organization.

      In the context of consolidated billing, you might see the term Payer Account. The Payer Account is the AWS account that contains the payment method for the organization. So, if you see Master Account, Management Account, or Payer Account, know that within AWS Organizations, they all refer to the same thing: the account used to create the organization and the account that contains the payment method for all accounts within the AWS Organization.

      Using consolidated billing within an AWS Organization means you receive a single monthly bill contained within the Management Account. This bill covers the Management Account and all Member Accounts of the organization. One bill contains all the billable usage for all accounts within the AWS Organization, removing a significant amount of financial admin overhead for larger businesses. This alone would be worth creating an organization for most larger enterprises.

      But it gets better. With AWS, certain services become cheaper the more you use them, and for certain services, you can pay in advance for cheaper rates. When using Organizations, these benefits are pooled, allowing the organization to benefit as a whole from the spending of each AWS account within it.

      AWS Organizations also features a service called Service Control Policies (SCPs), which allows you to restrict what AWS accounts within the organization can do. These are important, and I’ll cover them in their own dedicated lesson, which is coming up soon. I wanted to mention them now as a feature of AWS Organizations.

      Before we go through a demo where we'll create an AWS Organization and set up the final account structure for this course, I want to cover two other concepts. You can invite existing accounts into an organization, but you can also create new accounts directly within it. All you need is a valid, unique email address for the new account, and AWS will handle the rest. Creating accounts directly within the organization avoids the invite process required for existing accounts.

      Using an AWS Organization changes what is best practice in terms of user logins and permissions. With Organizations, you don’t need to have IAM Users inside every single AWS account. Instead, IAM roles can be used to allow IAM Users to access other AWS accounts. We’ll implement this in the following demo lesson. Best practice is to have a single account for logging in, which I’ve shown in this diagram as the Management Account of the organization. Larger enterprises might keep the Management Account clean and have a separate account dedicated to handling logins.

      Both approaches are fine, but be aware that the architectural pattern is to have a single AWS account that contains all identities for logging in. Larger enterprises might also have their own existing identity system and may use Identity Federation to access this single identity account. You can either use internal AWS identities with IAM or configure AWS to allow Identity Federation so that your on-premises identities can access this designated login account.

      From there, we can use this account with these identities and utilize a feature called role switching. Role switching allows users to switch roles from this account into other Member Accounts of the organization. This process assumes roles in these other AWS accounts. It can be done from the console UI, hiding much of the technical complexity, but it’s important to understand how it works. Essentially, you either log in directly to this login account using IAM identities or use Identity Federation to gain access to it, and then role switch into other accounts within the organization.

      I’ll discuss this in-depth as we progress through the course. The next lesson is a demo where you’ll implement this yourself and create the final AWS account structure for the remainder of the course.

      Okay, so at this point, it's time for a demo. As I mentioned, you'll be creating the account structure you'll use for the rest of the course. At the start, I demoed creating AWS accounts, including a general AWS account and a production AWS account. In the next lesson, I’ll walk you through creating an AWS Organization using this general account, which will become the Management Account for the AWS Organization. Then, you'll invite the existing production account into the organization, making it a Member Account. Finally, you'll create a new account within the organization, which will be the Development Account.

      I’m excited for this, and it’s going to be both fun and useful for the exam. So, go ahead and finish this video, and when you're ready, I look forward to you joining me in the next lesson, which will be a demo.

    1. Welcome back.

      In this lesson, I want to continue immediately from the last one by discussing when and where you might use IAM roles. By talking through some good scenarios for using roles, I want to make sure that you're comfortable with selecting these types of situations where you would choose to use an IAM role and where you wouldn't, because that's essential for real-world AWS usage and for answering exam questions correctly.

      So let's get started.

      One of the most common uses of roles within the same AWS account is for AWS services themselves. AWS services operate on your behalf and need access rights to perform certain actions. An example of this is AWS Lambda. Now, I know I haven't covered Lambda yet, but it's a function as a service product. What this means is that you give Lambda some code and create a Lambda function. This function, when it runs, might do things like start and stop EC2 instances, perform backups, or run real-time data processing. What it does exactly isn't all that relevant for this lesson. The key thing, though, is that a Lambda function, as with most AWS things, has no permissions by default. A Lambda function is not an AWS identity. It's a component of a service, and so it needs some way of getting permissions to do things when it runs. Running a Lambda function is known as a function invocation or a function execution using Lambda terminology.

      So anything that's not an AWS identity, this might be an application or a script running on a piece of compute hardware somewhere, needs to be given permissions on AWS using access keys. Rather than hard-coding some access keys into your Lambda function, there's actually a better way. To provide these permissions, we can create an IAM role known as a Lambda execution role. This execution role has a trust policy which trusts the Lambda service. This means that Lambda is allowed to assume that role whenever a function is executed. This role has a permissions policy which grants access to AWS products and services.

      When the function runs, it uses the sts:AssumeRole operation, and then the Secure Token Service generates temporary security credentials. These temporary credentials are used by the runtime environment in which the Lambda function runs to access AWS resources based on the permissions the role’s permissions policy has. The code is running in a runtime environment, and it's the runtime environment that assumes the role. The runtime environment gets these temporary security credentials, and then the whole environment, which the code is running inside, can use these credentials to access AWS resources.

      So why would you use a role for this? What makes this scenario perfect for using a role? Well, if we didn't use a role, you would need to hard-code permissions into the Lambda function by explicitly providing access keys for that function to use. Where possible, you should avoid doing that because, A, it's a security risk, and B, it causes problems if you ever need to change or rotate those access keys. It's always better for AWS products and services, where possible, to use a role, because when a role is assumed, it provides a temporary set of credentials with enough time to complete a task, and then these are discarded.

      For a given Lambda function, you might have one copy running at once, zero copies, 50 copies, a hundred copies, or even more. Because you can't determine this number, because it's unknown, if you remember my rule from the previous lesson, if you don't know the number of principals, if it's multiple or if it's an uncertain number, then it suggests a role might be the most ideal identity to use. In this case, it is the ideal way of providing Lambda with these credentials to use a role and allow it to get these temporary credentials. It's always the preferred option when using AWS services to do something on your behalf; use a role because you don't need to provide any static credentials.

      Okay, so let's move on to the next scenario.

      Another situation where roles are useful is emergency or out-of-the-usual situations. Here’s a familiar scenario that you might find in a workplace. This is Wayne, and Wayne works in a business's service desk team. This team is given read-only access to a customer's AWS account so that they can keep an eye on performance. The idea is that anything more risky than this read-only level of access is handled by a more senior technical team. We don't want to give Wayne's team long-term permissions to do anything more destructive than this read-only access, but there are always going to be situations which occur when we least want them, normally 3:00 a.m. on a Sunday morning, when a customer might call with an urgent issue where they need Wayne's help to maybe stop or start an instance, or maybe even terminate an EC2 instance and recreate it.

      So 99% of the time, Wayne and his team are happy with this read-only access, but there are situations when he needs more. This is a break-glass style situation, which is named after this. The idea of break glass in the physical world is that there is a key for something behind glass. It might be a key for a room that a certain team doesn't normally have access to, maybe it’s a safe or a filing cabinet. Whatever it is, the glass provides a barrier, meaning that when people break it, they really mean to break it. It’s a confirmation step. So if you break a piece of glass to get a key to do something, there needs to be an intention behind it. Anyone can break the glass and retrieve the key, but having the glass results in the action only happening when it's really needed. At other times, whatever the key is for remains locked. And you can also tell when it’s been used and when it hasn’t.

      A role can perform the same thing inside an AWS account. Wayne can assume an emergency role when absolutely required. When he does, he'll gain additional permissions based on the role's permissions policy. For a short time, Wayne will, in effect, become the role. This access will be logged and Wayne will know to only use the role under exceptional circumstances. Wayne’s normal permissions can remain at read-only, which protects him and the customer, but he can obtain more if required when it’s really needed. So that’s another situation where a role might be a great solution.

      Another scenario when roles come in handy is when you're adding AWS into an existing corporate environment. You might have an existing physical network and an existing provider of identities, known as an identity provider, that your staff use to log into various systems. For the sake of this example, let’s just say that it's Microsoft Active Directory. In this scenario, you might want to offer your staff single sign-on, known as SSO, allowing them to use their existing logins to access AWS. Or you might have upwards of 5,000 accounts. Remember, there’s the 5,000 IAM user limit. So for a corporation with more than 5,000 staff, you can’t offer each of them an IAM user. That is beyond the capabilities of IAM.

      Roles are often used when you want to reuse your existing identities for use within AWS. Why? Because external accounts can’t be used directly. You can’t access an S3 bucket directly using an Active Directory account. Remember this fact. External accounts or external identities cannot be used directly to access AWS resources. You can’t directly use Facebook, Twitter, or Google identities to interact with AWS. There is a separate process which allows you to use these external identities, which I’ll be talking about later in the course.

      Architecturally, what happens is you allow an IAM role inside your AWS account to be assumed by one of the external identities, which is in Active Directory in this case. When the role is assumed, temporary credentials are generated and these are used to access the resources. There are ways that this is hidden behind the console UI so that it appears seamless, but that's what happens behind the scenes. I'll be covering this in much more detail later in the course when I talk about identity federation, but I wanted to introduce it here because it is one of the major use cases for IAM roles.

      Now, why roles are so important when an existing ID provider such as Active Directory is involved is that, remember, there is this 5,000 IAM user limit in an account. So if your business has more than 5,000 accounts, then you can’t simply create an IAM user for each of those accounts, even if you wanted to. 5,000 is a hard limit. It can't be changed. Even if you could create more than 5,000 IAM users, would you actually want to manage 5,000 extra accounts? Using a role in this way, so giving permissions to an external identity provider and allowing external identities to assume this role, is called ID Federation. It means you have a small number of roles to manage and external identities can use these roles to access your AWS resources.

      Another common situation where you might use roles is if you're designing the architecture for a popular mobile application. Maybe it's a ride-sharing application which has millions of users. The application needs to store and retrieve data from a database product in AWS, such as DynamoDB. Now, I've already explained two very important but related concepts on the previous screen. Firstly, that when you interact with AWS resources, you need to use an AWS identity. And then secondly, that there’s this 5,000 IAM user limit per account. So designing an application with this many users which needs access to AWS resources, if you could only use IAM users or identities in AWS, it would be a problem because of this 5,000 user limit. It’s a hard limit and it can’t be raised.

      Now, this is a problem which can be fixed with a process called Web Identity Federation, which uses IAM roles. Most mobile applications that you’ve used, you might have noticed they allow you to sign in using a web identity. This might be Twitter, Facebook, Google, and potentially many others. If we utilize this architecture for our web application, we can trust these identities and allow these identities to assume an IAM role. This is based on that role’s trust policy. So they can assume that role, gain access to temporary security credentials, and use those credentials to access AWS resources, such as DynamoDB. This is a form of Web Identity Federation, and I'll be covering it in much more detail later in the course.

      The use of roles in this situation has many advantages. First, there are no AWS credentials stored in the application, which makes it a much more preferred option from a security point of view. If an application is exploited for whatever reason, there’s no chance of credentials being leaked, and it uses an IAM role which you can directly control from your AWS account. Secondly, it makes use of existing accounts that your customers probably already have, so they don't need yet another account to access your service. And lastly, it can scale to hundreds of millions of users and beyond. It means you don’t need to worry about the 5,000 user IAM limit. This is really important for the exam. There are very often questions on how you can architect solutions which will work for mobile applications. Using ID Federation, so using IAM roles, is how you can accomplish that. And again, I'll be providing much more information on ID Federation later in the course.

      Now, one scenario I want to cover before we finish up this lesson is cross-account access. In an upcoming lesson, I’ll be introducing AWS Organizations and you will get to see this type of usage in practice. It’s actually how we work in a multi-account environment. Picture the scenario that's on screen now: two AWS accounts, yours and a partner account. Let’s say your partner organization offers an application which processes scientific data and they want you to store any data inside an S3 bucket that’s in their account. Your account has thousands of identities, and the partner IT team doesn’t want to create IAM users in their account for all of your staff. In this situation, the best approach is to use a role in the partner account. Your users can assume that role, get temporary security credentials, and use those to upload objects. Because the IAM role in the partner account is an identity in that account, using that role means that any objects that you upload to that bucket are owned by the partner account. So it’s a very simple way of handling permissions when operating between accounts.

      Roles can be used cross-account to give access to individual resources like S3 in the onscreen example, or you can use roles to give access to a whole account. You’ll see this in the upcoming AWS Organization demo lesson. In that lesson, we’re going to configure it so a role in all of the different AWS accounts that we’ll be using for this course can be assumed from the general account. It means you won’t need to log in to all of these different AWS accounts. It makes multi-account management really simple.

      I hope by this point you start to get a feel for when roles are used. Even if you’re a little vague, you will learn more as you go through the course. For now, just a basic understanding is enough. Roles are difficult to understand at first, so you’re doing well if you’re anything but confused at this point. I promise you, as we go through the course and you get more experience, it will become second nature.

      So at this point, that’s everything I wanted to cover. Thanks for watching. Go ahead and complete this video, and when you're ready, join me in the next lesson.

    1. Welcome back.

      Over the next two lessons, I'll be covering a topic which is usually one of the most difficult identity-related topics in AWS to understand, and that's IAM roles. In this lesson, I'll step through how roles work, their architecture, and how you technically use a role. In the following lesson, I'll compare roles to IAM users and go into a little bit more detail on when you generally use a role, so some good scenarios which fit using an IAM role. My recommendation is that you watch both these lessons back to back in order to fully understand IAM roles.

      So let's get started.

      A role is one type of identity which exists inside an AWS account. The other type, which we've already covered, are IAM users. Remember the term "principal" that I introduced in the previous few lessons? This is a physical person, application, device, or process which wants to authenticate with AWS. We defined authentication as proving to AWS that you are who you say you are. If you authenticate, and if you are authorized, you can then access one or more resources.

      I also previously mentioned that an IAM user is generally designed for situations where a single principal uses that IAM user. I’ve talked about the way that I decide if something should use an IAM user: if I can imagine a single thing—one person or one application—who uses an identity, then generally under most circumstances, I'd select to use an IAM user.

      IAM roles are also identities, but they're used much differently than IAM users. A role is generally best suited to be used by an unknown number or multiple principals, not just one. This might be multiple AWS users inside the same AWS account, or it could be humans, applications, or services inside or outside of your AWS account who make use of that role. If you can't identify the number of principals which use an identity, then it could be a candidate for an IAM role. Or if you have more than 5,000 principals, because of the number limit for IAM users, it could also be a candidate for an IAM role.

      Roles are also something which is generally used on a temporary basis. Something becomes that role for a short period of time and then stops. The role isn't something that represents you. A role is something which represents a level of access inside an AWS account. It's a thing that can be used, short term, by other identities. These identities assume the role for a short time, they become that role, they use the permissions that that role has, and then they stop being that role. It’s not like an IAM user, where you login and it’s a representation of you, long term. With a role, you essentially borrow the permissions for a short period of time.

      I want to make a point of stressing that distinction. If you're an external identity—like a mobile application, maybe—and you assume a role inside my AWS account, then you become that role and you gain access to any access rights that that role has for a short time. You essentially become an identity in my account for a short period of time.

      Now, this is the point where most people get a bit confused, and I was no different when I first learned about roles. What's the difference between logging into a user and assuming a role? In both cases, you get the access rights that that identity has.

      Before we get to the end of this pair of lessons, so this one and the next, I think it's gonna make a little bit more sense, and definitely, as you go through the course and get some practical exposure to roles, I know it's gonna become second nature.

      IAM users can have identity permissions policies attached to them, either inline JSON or via attached managed policies. We know now that these control what permissions the identity gets inside AWS. So whether these policies are inline or managed, they're properly referred to as permissions policies—policies which grant, so allow or deny, permissions to whatever they’re associated with.

      IAM roles have two types of policies which can be attached: the trust policy and the permissions policy. The trust policy controls which identities can assume that role. With the onscreen example, identity A is allowed to assume the role because identity A is allowed in the trust policy. Identity B is denied because that identity is not specified as being allowed to assume the role in the trust policy.

      The trust policy can reference different things. It can reference identities in the same account, so other IAM users, other roles, and even AWS services such as EC2. A trust policy can also reference identities in other AWS accounts. As you'll learn later in the course, it can even allow anonymous usage of that role and other types of identities, such as Facebook, Twitter, and Google.

      If a role gets assumed by something which is allowed to assume it, then AWS generates temporary security credentials and these are made available to the identity which assumed the role. Temporary credentials are very much like access keys, which I covered earlier in the course, but instead of being long-term, they're time-limited. They only work for a certain period of time before they expire. Once they expire, the identity will need to renew them by reassuming the role, and at that point, new credentials are generated and given to the identity again which assumed that role.

      These temporary credentials will be able to access whatever AWS resources are specified within the permissions policy. Every time the temporary credentials are used, the access is checked against this permissions policy. If you change the permissions policy, the permissions of those temporary credentials also change.

      Roles are real identities and, just like IAM users, roles can be referenced within resource policies. So if a role can access an S3 bucket because a resource policy allows it or because the role permissions policy allows it, then anything which successfully assumes the role can also access that resource.

      You’ll get a chance to use roles later in this section when we talk about AWS Organizations. We’re going to take all the AWS accounts that we’ve created so far and join them into a single organization, which is AWS’s multi-account management product. Roles are used within AWS Organizations to allow us to log in to one account in the organization and access different accounts without having to log in again. They become really useful when managing a large number of accounts.

      When you assume a role, temporary credentials are generated by an AWS service called STS, or the Secure Token Service. This is the operation that's used to assume the role and get the credentials, so sts .

      In this lesson, I focused on the technical aspect of roles—mainly how they work. I’ve talked about the trust policy, the permissions policy, and how, when you assume a role, you get temporary security credentials. In the next lesson, I want to step through some example scenarios of where roles are used, and I hope by the end of that, you’re gonna be clearer on when you should and shouldn’t use roles.

      So go ahead, finish up this video, and when you’re ready, you can join me in the next lesson.

    1. Welcome back and welcome to this demo of the functionality provided by IAM Groups.

      What we're going to do in this demo is use the same architecture that we had in the IAM users demo—the SALLI user and those two S3 buckets—but we’re going to migrate the permissions that the SALLI user has from the user to a group that SALLI is a member of.

      Before we get started, just make sure that you are logged in as the IAM admin user of the general AWS account. As always, you’ll need to have the Northern Virginia region selected.

      Attached to this video is a demo files link that will download all of the files you’re going to use throughout the demo. To save some time, go ahead and click on that link and start the file downloading. Once it’s finished, go ahead and extract it; it will create a folder containing all of the files you’ll need as you move through the demo.

      You should have deleted all of the infrastructure that you used in the previous demo lesson. So at this point, we need to go ahead and recreate it. To do that, attached to this lesson is a one-click deployment link. So go ahead and click that link. Everything is pre-populated, so you need to make sure that you put in a suitable password that doesn’t breach any password policy on your account. I’ve included a suitable default password with some substitutions, so that should be okay for all common password policies.

      Scroll down to the bottom, click on the capabilities checkbox, and then create the stack. That’ll take a few moments to create, so I’m going to pause the video and resume it once that stack creation has completed.

      Okay, so that’s created now. Click on Services and open the S3 console in a new tab. This can be a normal tab. Go to the Cat Pics bucket, click Upload, add file, locate the demo files folder that you downloaded and extracted earlier. Inside that folder should be a folder called Cat Pics. Go in there and then select merlin.jpg. Click on Open and Upload. Wait for that to finish.

      Once it’s finished, go back to the console, go to Animal Pics, click Upload again, add files. This time, inside the Animal Pics folder, upload thaw.jpg. Click Upload. Once that’s done, go back to CloudFormation, click on Resources, and click on the Sally user. Inside the Sally user, click on Add Permissions, Attach Policies Directly, select the "Allow all S3 except cats" policy, click on Next, and then Add Permissions.

      So that brings us to the point where we were in the IAM users demo lesson. That’s the infrastructure set back up in exactly the same way as we left the IAM users demo. Now we can click on Dashboard. You’ll need to copy the IAM signing users link for the general account. Copy that into your clipboard.

      You’re going to need a separate browser, ideally, a fully separate browser. Alternatively, you can use a private browsing tab in your current browser, but it’s just easier to understand probably for you at this point in your learning if you have a separate browser window. I’m going to use an isolated tab because it’s easier for me to show you.

      You’ll need to paste in this IAM URL because now we’re going to sign into this account using the Sally user. Go back to CloudFormation, click on Outputs, and you’ll need the Sally username. Copy that into your clipboard. Go back to this separate browser window and paste that in. Then, back to CloudFormation, go to the Parameters tab and get the password for the Sally user. Enter the password that you chose for Sally when you created the stack.

      Then move across to the S3 console and just verify that the Sally user has access to both of these buckets. The easiest way of doing that is to open both of these animal pictures. We’ll start with Thor. Thor’s a big doggo, so it might take some time for him to load in. There we go, he’s loaded in. And the Cat Pics bucket. We get access denied because remember, Sally doesn’t have access to the Cat Pics bucket. That’s as intended.

      Now we’ll go back to our other browser window—the one where we logged into the general account as the IAM admin user. This is where we’re going to make the modifications to the permissions. We’re going to change the permissions over to using a group rather than directly on the Sally user.

      Click on the Resources tab first and select Sally to move across to the Sally user. Note how Sally currently has this managed policy directly attached to her user. Step one is to remove that. So remove this managed policy from Sally. Detach it. This now means that Sally has no permissions on S3. If we go back to the separate browser window where we’ve got Sally logged in and then hit refresh, we see she doesn’t have any permissions now on S3.

      Now back to the other browser, back to the one where we logged in as IAM admin, click on User Groups. We’re going to create a Developers group. Click on Create New Group and call it Developers. That’s the group name. Then, down at the bottom here, this is where we can attach a managed policy to this group. We’re going to attach the same managed policy that Sally had previously directly on her user—Allow all S3 except cats.

      Type "allow" into the filter box and press Enter. Then check the box to select this managed policy. We could also directly at this stage add users to this group, but we’re not going to do that. We’re going to do that as a separate process. So click on 'Create Group'.

      So that’s the Developers group created. Notice how there are not that many steps to create a group, simply because it doesn’t offer that much in the way of functionality. Open up the group. The only options you see here are 'User Membership' and any attached permissions. Now, as with a user, you can attach inline policies or managed policies, and we’ve got the managed policy.

      What we’re going to do next is click on Users and then Add Users to Group. We’re going to select the Sally IAM user and click on Add User. Now our IAM user Sally is a member of the Developers group, and the Developers group has this attached managed policy that allows them to access everything on S3 except the Cat Pics bucket.

      Now if I move back to my other browser window where I’ve got the Sally user logged in and then refresh, now that the Sally user has been added to that group, we’ve got permissions again over S3. If I try to access the Cat Pics bucket, I won’t be able to because that managed policy that the Developers team has doesn’t include access for this. But if I open the Animal Pics bucket and open Thor again—Thor’s a big doggo, so it’ll take a couple of seconds—it will load in that picture absolutely fine.

      So there we go, there’s Thor. That’s pretty much everything I wanted to demonstrate in this lesson. It’s been a nice, quick demo lesson. All we’ve done is create a new group called Developers, added Sally to this Developers group, removed the managed policy giving access to S3 from Sally directly, and added it to the Developers group that she’s now a member of. Note that no matter whether the policy is attached to Sally directly or attached to a group that Sally is a member of, she still gets those permissions.

      That’s everything I wanted to cover in this demo lesson. So before we finish up, let’s just tidy up our account. Go to Developers and then detach this managed policy from the Developers group. Detach it, then go to Groups and delete the Developers group because it wasn’t created as part of the CloudFormation template.

      Then, as the IAM admin user, open up the S3 console. We need to empty both of these buckets. Select Cat Pics, click on Empty. You’ll need to type or copy and paste 'Permanently Delete' into that box and confirm the deletion. Click Exit. Then select the Animal Pics bucket and do the same process. Copy and paste 'Permanently Delete' and confirm by clicking on Empty and then Exit.

      Now that we’ve done that, we should have no problems opening up CloudFormation, selecting the IAM stack, and then hitting Delete. Note if you do have any errors deleting this stack, just go into the stack, select Events, and see what the status reason is for any of those deletion problems. It should be fairly obvious if it can’t delete the stack because it can’t delete one or more resources, and it will give you the reason why.

      That being said, at this point, assume the stack deletions worked successfully, and we’ve cleaned up our account. That’s everything I wanted to cover in this demo lesson. Go ahead, complete this video, and when you’re ready, I’ll see you in the next lesson.

    1. Welcome back.

      In this lesson, I want to briefly cover IAM groups, so let's get started.

      IAM groups, simply put, are containers for IAM users. They exist to make organizing large sets of IAM users easier. You can't log in to IAM groups, and IAM groups have no credentials of their own. The exam might try to trick you on this one, so it's definitely important that you remember you cannot log into a group. If a question or answer suggests logging into a group, it's just simply wrong. IAM groups have no credentials, and you cannot log into them. So they're used solely for organizing IAM users to make management of IAM users easier.

      So let's look at a visual example. We've got an AWS account, and inside it we've got two groups: Developers and QA. In the Developers group, we've got Sally and Mike. In the QA group, we've got Nathalie and Sally. Now, the Sally user—so the Sally in Developers and the Sally in the QA group—that's the same IAM user. An IAM user can be a member of multiple IAM groups. So that's important to remember for the exam.

      Groups give us two main benefits. First, they allow effective administration-style management of users. We can make groups that represent teams, projects, or any other functional groups inside a business and put IAM users into those groups. This helps us organize.

      Now, the second benefit, which builds off the first, is that groups can actually have policies attached to them. This includes both inline policies and managed policies. In the example on the screen now, the Developers group has a policy attached, as does the QA group. There’s also nothing to stop IAM users, who are themselves within groups, from having their own inline or managed policies. This is the case with Sally.

      When an IAM user such as Sally is added as a member of a group—let’s say the Developers group—that user gets the policies attached to that group. Sally gains the permissions of any policies attached to the Developers group and any other groups that that user is a member of. So Sally also gets the policies attached to the QA group, and Sally has any policies that she has directly.

      With this example, Sally is a member of the Developers group, which has one policy attached, a member of the QA group with an additional policy attached, and she has her own policy. AWS merges all of those into a set of permissions. So effectively, she has three policies associated with her user: one directly, and one from each of the group memberships that her user has.

      When you're thinking about the allow or deny permissions in policy statements for users that are in groups, you need to consider those which apply directly to the user and their group memberships. Collect all of the policy allows and denies that a user has directly and from their groups, and apply the same deny-allow-deny rule to them as a collection. Evaluating whether you're allowed or denied access to a resource doesn’t become any more complicated; it’s just that the source of those allows and denies can broaden when you have users that are in multiple IAM groups.

      I mentioned last lesson that an IAM user can be a member of up to 10 groups and there is a 5,000 IAM user limit for an account. Neither of those are changeable; they are hard limits. There’s no effective limit for the number of users in a single IAM group, so you could have all 5,000 IAM users in an account as members of a single IAM group.

      Another common area of trick questions in the exam is around the concept of an all-users group. There isn't actually a built-in all-users group inside IAM, so you don’t have a single group that contains all of the members of that account like you do with some other identity management solutions. In IAM, you could create a group and add all of the users in that account into the group, but you would need to create and manage it yourself. So that doesn’t exist natively.

      Another really important limitation of groups is that you can’t have any nesting. You can’t have groups within groups. IAM groups contain users and IAM groups can have permissions attached. That’s it. There’s no nesting, and groups cannot be logged into; they don’t have any credentials.

      Now, there is a limit of 300 groups per account, but this can be increased with a support ticket.

      There’s also one more point that I want to make at this early stage in the course. This is something that many other courses tend to introduce later on or at a professional level, but it's important that you understand this from the very start. I'll show you later in the course how policies can be attached to resources, for example, S3 buckets. These policies, known as resource policies, can reference identities. For example, a bucket could have a policy associated with it that allows Sally access to that bucket. That’s a resource policy. It controls access to a specific resource and allows or denies identities to access that bucket.

      It does this by referencing these identities using an ARN, or Amazon Resource Name. Users and IAM roles, which I'll be talking about later in the course, can be referenced in this way. So a policy on a resource can reference IAM users and IAM roles by using the ARN. A bucket could give access to one or more users or to one or more roles, but groups are not a true identity. They can’t be referenced as a principal in a policy. A resource policy cannot grant access to an IAM group. You can grant access to IAM users, and those users can be in groups, but a resource policy cannot grant access to an IAM group. It can’t be referred to in this way. You couldn’t have a resource policy on an S3 bucket and grant access to the Developers group and then expect all of the developers to access it. That’s not how groups work. Groups are just there to group up IAM users and allow permissions to be assigned to those groups, which the IAM users inherit.

      So this is an important one to remember, whether you are answering an exam question that involves groups, users, and roles or resource policies, or whether you're implementing real-world solutions. It’s easy to overestimate the features that a group provides. Don’t fall into the trap of thinking that a group offers more functionality than it does. It’s simply a container for IAM users. That’s all it’s for. It can contain IAM users and have permissions associated with it; that’s it. You can’t log in to them and you can’t reference them from resource policies.

      Okay, so that’s everything I wanted to cover in this lesson. Go ahead, complete the video, and when you're ready, I'll look forward to you joining me in the next.

    1. Welcome back.

      And in this lesson, I want to finish my coverage of IAM users.

      You already gained some exposure to IAM users earlier in the course. Remember, you created an IAM admin user in both your general and production AWS accounts. As well as creating these users, you secured them using MFA, and you attached an AWS managed policy to give this IAM user admin rights in both of those accounts.

      So for now, I just want to build upon your knowledge of IAM users by adding some extra detail that you'll need for the exam. So let's get started.

      Now, before I go into more detail, let's just establish a foundation. Let's use a definition. Simply put, IAM users are an identity used for anything requiring long-term AWS access. For example, humans, applications, or service accounts. If you need to give something access to your AWS account, and if you can picture one thing, one person or one application—so James from accounts, Mike from architecture, or Miles from development—99% of the time you would use an IAM user.

      If you need to give an application access to your AWS account, for example, a backup application running on people's laptops, then each laptop generally would use an IAM user. If you have a need for a service account, generally a service account which needs to access AWS, then generally this will use an IAM user. If you can picture one thing, a named thing, then 99% of the time, the correct identity to select is an IAM user. And remember this because it will help in the exam.

      IAM starts with a principal. And this is a word which represents an entity trying to access an AWS account. At this point, it's unidentified. Principals can be individual people, computers, services, or a group of any of those things. For a principal to be able to do anything, it needs to authenticate and be authorized. And that's the process that I want to step through now.

      A principal, which in this example, is a person or an application, makes requests to IAM to interact with resources. Now, to be able to interact with resources, it needs to authenticate against an identity within IAM. An IAM user is an identity which can be used in this way.

      Authentication is this first step. Authentication is a process where the principal on the left proves to IAM that it is an identity that it claims to be. So an example of this is that the principal on the left might claim to be Sally, and before it can use AWS, it needs to prove that it is indeed Sally. And it does this by authenticating.

      Authentication for IAM users is done either using username and password or access keys. These are both examples of long-term credentials. Generally, username and passwords are used if a human is accessing AWS and accessing via the console UI. Access keys are used if it's an application, or as you experienced earlier in the course, if it's a human attempting to use the AWS Command Line tools.

      Now, once a principal goes through the authentication process, the principal is now known as an authenticated identity. An authenticated identity has been able to prove to AWS that it is indeed the identity that it claims to be. So it needs to be able to prove that it's Sally. And to prove that it's Sally, it needs to provide Sally's username and password, or be able to use Sally's secret access key, which is a component of the access key set. If it can do that, then AWS will know that it is the identity that it claims to be, and so it can start interacting with AWS.

      Once the principal becomes an authenticated identity, then AWS knows which policies apply to the identity. So in the previous lesson, I talked about policy documents, how they could have one or more statements, and if an identity attempted to access AWS resources, then AWS would know which statements apply to that identity. That's the process of authorization.

      So once a principal becomes an authenticated identity, and once that authenticated identity tries to upload to an S3 bucket or terminate an EC2 instance, then AWS checks that that identity is authorized to do so. And that's the process of authorization. So they're two very distinct things. Authentication is how a principal can prove to IAM that it is the identity that it claims to be using username and password or access keys, and authorization is IAM checking the statements that apply to that identity and either allowing or denying that access.

      Okay, let's move on to the next thing that I want to talk about, which is Amazon Resource Names, or ARNs. ARNs do one thing, and that's to uniquely identify resources within any AWS accounts. When you're working with resources, using the command line or APIs, you need a way to refer to these resources in an unambiguous way. ARNs allow you to refer to a single resource, if needed, or in some cases, a group of resources using wild cards.

      Now, this is required because things can be named in a similar way. You might have an EC2 instance in your account with similar characteristics to one in my account, or you might have two instances in your account but in different regions with similar characteristics. ARNs can always identify single resources, whether they're individual resources in the same account or in different accounts.

      Now, ARNs are used in IAM policies which are generally attached to identities, such as IAM users, and they have a defined format. Now, there are some slight differences depending on the service, but as you go through this course, you'll gain enough exposure to be able to confidently answer any exam questions that involve ARNs. So don't worry about memorizing the format at this stage, you will gain plenty of experience as we go.

      These are two similar, yet very different ARNs. They both look to identify something related to the catgifs bucket. They specify the S3 service. They don't need to specify a region or an account because the naming of S3 is globally unique. If I use a bucket name, then nobody else can use that bucket name in any account worldwide.

      The difference between these two ARNs is the forward slash star on the end at the second one. And this difference is one of the most common ways mistakes can be made inside policies. It trips up almost all architects or admins at one point or another. The top ARN references an actual bucket. If you wanted to allow or deny access to a bucket or any actions on that bucket, then you would use this ARN which refers to the bucket itself. But a bucket and objects in that bucket are not the same thing.

      This ARN references anything in that bucket, but not the bucket itself. So by specifying forward slash star, that's a wild card that matches any keys in that bucket, so any object names in that bucket. This is really important. These two ARNs don't overlap. The top one refers to just the bucket and not the objects in the bucket. The bottom one refers to the objects in the bucket but not the bucket itself.

      Now, some actions that you want to allow or deny in a policy operate at a bucket level or actually create buckets. And this would need something like the top ARN. Some actions work on objects, so it needs something similar to the bottom ARN. And you need to make sure that you use the right one. In some cases, creating a policy that allows a set of actions will need both. If you want to allow access to create a bucket and interact with objects in that bucket, then you would potentially need both of these ARNs in a policy.

      ARNs are collections of fields split by a colon. And if you see a double colon, it means that nothing is between it. It doesn't need to be specified. So in this example, you'll see a number of double colons because you don't need to specify the region or account number for an S3 bucket because the bucket name is globally unique. A star can also be used, which is a wild card.

      Now, keep in mind they're not the same thing. So not specifying a region and specifying star don't mean the same thing. You might use a star when you want to refer to all regions inside an AWS account. Maybe you want to give permissions to interact with EC2 in all regions, but you can't simply omit this. The only place you'll generally use the double colon is when something doesn't need to be specified, you'd use a star when you want to refer to a wild card collection of a set of things. So they're not the same thing. Keep that in mind, and I'll give you plenty of examples as we go through the course.

      So the first field is the partition, and this is the partition that the resource is in. For standard AWS regions, the partition is AWS. If you have resources in other partitions, the partition is AWS-hyphen-partition name. This is almost never anything but AWS. But for example, if you do have resources in the China Beijing region, then this is AWS-cn.

      The next part is service. And this is the service name space that identifies the AWS product. For example, S3, IAM, or RDS. The next field is region. So this is the region that the resource you're referring to resides in. Some ARNs do not require a region, so this might be omitted, and certain ARNs require wild card. And you'll gain exposure through the course as to what different services require for their ARNs.

      The next field is the account ID. This is the account ID of the AWS account that owns the resource. So for example, 123456789012. So if you're referring to an EC2 instance in a certain account, you will have to specify the account number inside the ARN. Some resources don't require that, so this example is S3 because it is globally unique across every AWS account. You don't need to specify the account number.

      And then at the end, we've got resource or resource type. And the content of this part of the ARN varies depending on the service. A resource identifier can be the name or ID of an object. For example, user forward slash Sally or instance forward slash and then the instance ID, or it can be a resource path. But again, I'm only introducing this at this point. You'll get plenty of exposure as you go through the course. I just want to give you this advanced knowledge so you know what to expect.

      So let's quickly talk about an exam PowerUp. I tend not to include useless facts and figures in my course, but some of them are important. This is one such occasion.

      Now first, you can only ever have 5,000 IAM users in a single account. IAM is a global service, so this is a per account limit, not per region. And second, an IAM user can be a member of 10 IAM groups. So that's a maximum. Now, both of these have design impacts. You need to be aware of that.

      What it means is that if you have a system which requires more than 5,000 identities, then you can't use one IAM user for each identity. So this might be a limit for internet scale applications with millions of users, or it might be a limit for large organizations which have more than 5,000 staff, or it might be a limit when large organizations are merging together. If you have any scenario or a project with more than 5,000 identifiable users, so identities, then it's likely that IAM users are not the right identity to pick for that solution.

      Now, there are solutions which fix this. We can use IAM roles or Identity Federation, and I'll be talking about both of those later in the course. But in summary, it means using your own existing identities rather than using IAM users. And I'll be covering the architecture and the implementation of this later in the course.

      At this stage, I want you to take away one key fact, and that is this 5,000 user limit. If you are faced with an exam question which mentions more than 5,000 users, or talks about an application that's used on the internet which could have millions of users, and if you see an answer saying create an IAM user for every user of that application, that is the wrong answer. Generally with internet scale applications, or enterprise access or company mergers, you'll be using Federation or IAM roles. And I'll be talking about all of that later in the course.

      Okay, so that's everything I wanted to cover in this lesson. So go ahead, complete the video, and when you're ready, I'll look forward to you joining me in the next.

    1. Welcome back. In this lesson, I want to start by covering an important aspect of how AWS handles security, specifically focusing on IAM policies.

      IAM policies are a type of policy that gets attached to identities within AWS. As you've previously learned, identities include IAM users, IAM groups, and IAM roles. You’ll use IAM policies frequently, so it’s important to understand them for the exam and for designing and implementing solutions in AWS.

      Policies, once you understand them, are actually quite simple. I’ll walk you through the components and give you an opportunity to experiment with them in your own AWS account. Understanding policies involves three main stages: first, understanding their architecture and how they work; second, gaining the ability to read and understand the policy; and finally, learning to write your own. For the exam, understanding their architecture and being able to read them is sufficient. Writing policies will come as you work through the course and gain more practical experience.

      Let's jump in. An IAM identity policy, or IAM policy, is essentially a set of security statements for AWS. It grants or denies access to AWS products and features for any identity using that policy. Identity policies, also known as policy documents, are created using JSON. Familiarity with JSON is helpful, but if you're new to it, don’t worry—it just requires a bit more effort to learn.

      This is an example of an identity policy document that you would use with a user, group, or role. At a high level, a policy document consists of one or more statements. Each statement is enclosed in curly braces and grants or denies permissions to AWS services.

      When an identity attempts to access AWS resources, it must prove its identity through a process known as authentication. Once authenticated, AWS knows which policies apply to that identity, and each policy can contain multiple statements. AWS also knows which resources you’re trying to interact with and what actions you want to perform on those resources. AWS reviews all relevant statements one by one to determine the permissions for a given identity accessing a particular resource.

      A statement consists of several parts. The first part is a statement ID, or SID, which is optional but helps identify the statement and its purpose. For example, "full access" or "DenyCatBucket" indicates what the statement does. Using these identifiers is considered best practice.

      Every interaction with AWS involves two main elements: the resource and the actions attempted on that resource. For instance, if you’re interacting with an S3 bucket and trying to add an object, the statement will only apply if it matches both the action and the resource. The action part of a statement specifies one or more actions, which can be very specific or use wildcards (e.g., s3:* for all S3 operations). Similarly, resources can be specified individually or in lists, and wildcards can refer to all resources.

      The final component is the effect, which is either "allow" or "deny." The effect determines what AWS does if the action and resource parts of the statement match the attempted operation. If the effect is "allow," access is granted; if it’s "deny," access is blocked. An explicit deny always takes precedence over an explicit allow. If neither applies, the default implicit deny prevails.

      In scenarios where there are multiple policies or statements, AWS evaluates all applicable statements. If there’s an explicit deny, it overrides any explicit allows. If no explicit deny is present, an explicit allow will grant access, unless there’s an explicit deny.

      Lastly, there are two main types of policies: inline policies and managed policies. Inline policies are directly attached to individual identities, making them isolated and cumbersome to manage for large numbers of users. Managed policies are created as separate objects and can be attached to multiple identities, making them more efficient and easier to manage. AWS provides managed policies, but you can also create and manage customer managed policies tailored to your specific needs.

      Before concluding, you’ll have a chance to gain practical experience with these policies. For now, this introduction should give you a solid foundation. Complete the video, and I look forward to seeing you in the next lesson.

    1. The daily cards and journal entries are obviously indexed by chronological date and then within tabbed sections by month and year.

      The rest of the other cards with notes are given individual (decimal) numbers and and then are put into numerical order. These numbered cards are then indexed by putting related subject/topic/category words from them onto a separate index card which cross references either a dated card or the numbered card to which it corresponds. These index cards with topical words/phrases are then filed alphabetically into a tabbed alphabetical section (A-Z).

      As an example with the card in this post, if I wanted to remember all the books I buy from Octavia's Bookshelf, then I'd create a card titled "Octavia's Bookshelf" and list the title along with the date 2024-08-13 and file it alphabetically within the "O" tab section of the index. Obviously this might be more useful if I had more extensive notes about the book or its purchase on the 2024-08-13 card. I did create a short journal card entry about the bookstore on 08-13 because it was the first time I visited the bookstore in it's new location and decor, so there are some scant notes about my impressions of that which are cross-indexed to that Octavia's Bookshelf card. Thus my Octavia's Bookshelf card has an entry with "The Book Title, 2024-08-13 (J)(R)" where the '(J)' indicates there's a separate journal entry for that day and the '(R)' indicates there's also a receipt filed next to that day's card.

      I also created an "Author Card" with the author of the book's name, the title, publication date, etc. I included the purchase date and the reason why I was interested in the book. I'll use that same card to write notes on that particular book as I read it. These author cards are filed in a separate A-Z tabbed 'Bibliography' section for easily finding them as well. (I suppose I could just put them into the primary A-Z index, but I prefer having all the authors/books (I have thousands) in the same section.)

      I also have a rolodex section of people filed alphabetically, so I can easily look up Steve and Sonia separately and see what I might have gotten them on prior birthdays as well as notes about potential future gift ideas. I had tickler cards with their names on them filed in early August and now that they're in my to do list, I've moved those cards to August 2025, ready for next year's reminder. Compared to a typical Future Log I don't do nearly as much writing and rewriting when migrating. I just migrate a card forward until it's done or I don't need it anymore.

      If you've used a library card index before, the general idea is roughly the same, you're just cross-indexing more than books by title, author and subject. You can index by day, idea, project, or any other thing you like. My card index cabinet is really just a large personal database made out of paper and metal.

      The secret isn't to index everything—just the things you either want to remember or know you'll want to look up later and use/re-use.

    1. Author response:

      The following is the authors’ response to the original reviews.

      eLife assessment

      In this useful study, the authors report the efficacy, hematological effects, and inflammatory response of the BPaL regimen (containing bedaquiline, pretomanid, and linezolid) compared to a variation in which Linezolid is replaced with the preclinical development candidate spectinamide 1599, administered by inhalation in tuberculosis-infected mice. The authors provide convincing evidence that supports the replacement of Linezolid in the current standard of care for drug-resistant tuberculosis. However, a limitation of the work is the lack of control experiments with bedaquiline and pretomanid only, to further dissect the relevant contributions of linezolid and spectinamide in efficacy and adverse effects.

      We acknowledge a limitation in our study due to lack of groups with monotherapy of bedaquiline and pretomanid however, similar studies to understand contribution of bedaquiline and pretomanid to the BPaL have been published already (references #4 and #60 in revised manuscript).  Our goal was to compare the BPaS versus the BPaL with the understanding that TB treatment requires multidrug therapy.   We omitted monotherapy groups to reduce complexity of the studies because the multidrug groups require very large number of animals with very intensive and complex dosing schedules. Even if B or Pa by themselves have better efficacy than the BPa or BPaL combination, patients will not be treated with only B or Pa because of very high risk of developing drug resistance to B or/and PA. If drug resistance is developed for B or Pa, the field will lose very effective drugs against TB. 

      Although the manuscript is well written overall, a re-formulation of some of the stated hypotheses and conclusions, as well as the addition of text to contextualize translatability, would improve value.

      Manuscript has been edited to address these critiques.  Answers to individual critiques are below.

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      This manuscript is an extension of previous studies by this group looking at the new drug spectinamide 1599. The authors directly compare therapy with BPaL (bedaquiline, pretomanid, linezolid) to a therapy that substitutes spectinamide for linezolid (BPaS). The Spectinamide is given by aerosol exposure and the BPaS therapy is shown to be as effective as BPaL without adverse effects. The work is rigorously performed and analyses of the immune responses are consistent with curative therapy.

      Strengths:

      (1) This group uses 2 different mouse models to show the effectiveness of the BPaS treatment.

      (2) Impressively the group demonstrates immunological correlates associated with Mtb cure with the BPaS therapy.

      (3) Linezolid is known to inhibit ribosomes and mitochondria whereas spectinaminde does not. The authors clearly demonstrate the lack of adverse effects of BPaS compared to BPaL.

      Weaknesses:

      (1) Although this is not a weakness of this paper, a sentence describing how the spectinamide would be administered by aerosolization in humans would be welcomed.

      We already reported on the aerodynamic properties of dry powder spectinamide 1599 within #3 HPMC capsules and its delivery from a RS01 Plastiape inhaler device (reference #59 in revised manuscript).  To address this critique, we added a last paragraph in discussion “It is proposed that human use of spectinamides 1599 will be administered using a dry powder formulation delivered by the RS01 Plastiape dry powder inhaler" (reference #59 in revised manuscript).  

      Reviewer #2 (Public Review):

      Summary:

      Replacing linezolid (L) with the preclinical development candidate spectinamide 1599, administered by inhalation, in the BPaL standard of care regimen achieves similar efficacy, and reduces hematological changes and proinflammatory responses.

      Strengths:

      The authors not only measure efficacy but also quantify histological changes, hematological responses, and immune responses, to provide a comprehensive picture of treatment response and the benefits of the L to S substitution.

      The authors generate all data in two mouse models of TB infection, each reproducing different aspects of human histopathology.

      Extensive supplementary figures ensure transparency. 

      Weaknesses:

      The articulation of objectives and hypotheses could be improved.

      We edited to "The AEs were associated with the long-term administration of the protein synthesis inhibitor linezolid. Spectinamide 1599 (S) is also a protein synthesis inhibitor of Mycobacterium tuberculosis with an excellent safety profile, but which lacks oral bioavailability. Here, we propose to replace L in the BPaL regimen with spectinamide administered via inhalation and we demonstrate that inhaled spectinamide 1599, combined with BPa ––BPaS regimen––has similar efficacy to that of BPaL regimen while simultaneously avoiding the L-associated AEs.

      Reviewer #3 (Public Review):

      Summary:

      In this paper, the authors sought to evaluate whether the novel TB drug candidate, spectinamide 1599 (S), given via inhalation to mouse TB models, and combined with the drugs B (bedaquiline) and Pa (pretomanid), would demonstrate similar efficacy to that of BPaL regimen (where L is linezolid). Because L is associated with adverse events when given to patients long-term, and one of those is associated with myelosuppression (bone marrow toxicity) the authors also sought to assess blood parameters, effects on bone marrow, immune parameters/cell effects following treatment of mice with BPaS and BPaL. They conclude that BPaL and BPaS have equivalent efficacy in both TB models used and that BPaL resulted in weight loss and anemia (whereas BPaL did not) under the conditions tested, as well as effects on bone marrow.

      Strengths:

      The authors used two mouse models of TB that are representative of different aspects of TB in patients (which they describe well), intending to present a fuller picture of the activity of the tested drug combinations. They conducted a large body of work in these infected mice to evaluate efficacy and also to survey a wide range of parameters that could inform the effect of the treatments on bone marrow and on the immune system. The inclusion of BPa controls (in most studies) and also untreated groups led to a large amount of useful data that has been collected for the mouse models per se (untreated) as well as for BPa - in addition to the BPaS and BPaL combinations which are of particular interest to the authors. Many of these findings related to BPa, BPaL, untreated groups, etc corroborate earlier findings and the authors point this out effectively and clearly in their manuscript. To go further, in general, it is a well-written and cited article with an informative introduction.

      Weaknesses:

      The authors performed a large amount of work with the drugs given at the doses and dosing intervals started, but at present, there is no exposure data available in the paper. It would be of great value to understand the exposures achieved in plasma at least (and in the lung if more relevant for S) in order to better understand how these relate to clinical exposures that are observed at marketed doses for B, Pa, and L as well as to understand the exposure achieved at the doses being evaluated for S. If available as historical data this could be included/cited. Considering the great attempts made to evaluate parameters that are relevant to clinical adverse events, it would add value to understand what exposures of drug effects such as anemia, weight loss, and bone marrow effects, are being observed. It would also be of value to add an assessment of whether the weight loss, anemia, or bone marrow effects observed for BPaL are considered adverse, and the extent to which we can translate these effects from mouse to patient (i.e. what are the limitations of these assessments made in a mouse study?). For example, is the small weight loss seen as significant, or is it reversible? Is the magnitude of the changes in blood parameters similar to the parameters seen in patients given L? In addition, it is always challenging to interpret findings for combinations of drugs, so the addition of language to explain this would add value: for example, how confident can we be that the weight loss seen for only the BPaL group is due to L as opposed to a PK interaction leading to an elevated exposure and weight loss due to B or Pa?

      We totally agree with this critique but the studies suggested by the reviewer are very expensive and

      logistically/resource intensive. Data reported in this manuscript was used as preliminary data in a RO1 application to NIH-NIAID that included studies proposed above by this reviewer. The authors are glad to report that the application got a fundable score and is currently under consideration for funding by NIH-NIAID.   The summary of proposed future studies is included in the last paragraph of the discussion in this revised manuscript. 

      Turning to the evaluations of activity in mouse TB models, unfortunately, the evaluations of activity in the BALB/c mouse model as well as the spleens of the Kramnik model resulted in CFU below/at the limit of detection and so, to this reviewer's understanding of the data, comparisons between BPaL and BPaS cannot be made and so the conclusion of equivalent efficacy in BALB/c is not supported with the data shown. There is no BPa control in the BALB/c study, therefore it is not possible to discern whether L or S contributed to the activity of BPaL or BPaS; it is possible that BPa would have shown the same efficacy as the 3 drug combinations. It would be valuable to conduct a study including a BPa control and with a shorter treatment time to allow comparison of BPa, BPaS, and BPaL. 

      We agree with the reviewer these studies need to be done.  Some of them were recently published by our colleague Dr. Lyons (reference #60 in revised manuscript). The studies proposed by the reviewer will be performed under a new award under consideration for funding by the NIH-NIAID, the summary of future studies is included in the last paragraph of the discussion in this revised manuscript. 

      In the Kramnik lungs, as the authors rightly note, the studies do not support any contribution of S or L to BPa - i.e. the activity observed for BPa, BPaL, and BPaS did not significantly differ. Although the conclusions note equivalency of BPaL and BPaS, which is correct, it would be helpful to also include BPa in this statement;

      We edited and now included in lines #191 as requested 

      It would be useful to conduct a study dosing for a longer period of time or assessing a relapse endpoint, where it is possible that a contribution of L and/or S may be seen - thus making a stronger argument for S contributing an equivalent efficacy to L. The same is true for the assessment of lesions - unfortunately, there was no BPa control meaning that even where equivalency is seen for BPaL and BPaS, the reader is unable to deduce whether L or S made a contribution to this activity.

      Added in the future plans in the last paragraph of discussion

      “Future studies are already under consideration for funding by NIH-NIAID to understand the pharmacokinetics of mono, binary and ternary combinations of BPaS. These studies also aim to identify the optimal dose level and dosing frequency of each regimen along with their efficacy and relapse free-sterilization potential. Studies are also planned using a model-based pharmacokinetic-pharmacodynamic (PKPD) framework, guided by an existing human BPa PKPD model (reference #61 in revised manuscript), to find allometric human dose levels, dosing frequencies and treatment durations that will inform the experimental design of future clinical studies. 

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors):

      Although this is not a weakness of this paper, a sentence describing how the spectinamide would be administered by aerosolization in humans would be welcomed.

      Last paragraph of discussion was added “It is proposed that human use of spectinamides 1599 will be administered using a dry powder formulation delivered by the RS01 Plastiape dry powder inhaler". We already reported on the aerodynamic properties of dry powder spectinamide 1599 within #3 HPMC capsules and delivered from a RS01 Plastiape inhaler device (reference #59 in revised manuscript)

      Reviewer #2 (Recommendations For The Authors):

      Major comments

      The Abstract lacks focus and could more clearly convey the key messages.

      Edited as requested 

      The two mouse models and why they were chosen need to be described earlier. Currently, it's covered in the first section of the Discussion, but the reader needs to understand the utility of each model in answering the questions at hand before the first results are described, either in the introduction or in the opening section of the results.

      Thank you for suggestion, we agree.  We moved the first paragraph in discussion to last paragraph in Introduction. 

      Line 130: Please justify the doses and dosing frequency for S. A reference to a published manuscript could suffice if compelling.

      The dosing and regimens were previously reported by our groups in ref 21 and 22 in revised manuscript.- 

      (21) Robertson GT, Scherman MS, Bruhn DF, Liu J, Hastings C, McNeil MR, et al. Spectinamides are effective partner agents for the treatment of tuberculosis in multiple mouse infection models. J Antimicrob Chemother.

      2017;72(3):770–7. 

      (22) Gonzalez-Juarrero M, Lukka PB, Wagh S, Walz A, Arab J, Pearce C, et al. Preclinical Evaluation of Inhalational Spectinamide-1599 Therapy against Tuberculosis. ACS Infect Dis. 2021;7(10):2850–63. 

      Figures 1 E to H: several "ns" are missing, please add them.

      Edited as requested 

      Line 184 to 190: suggest moving the body weight plots to a Supplemental Figure, and at least double the size of the histology images to convey the message of lines 192-203.

      Please include higher magnification insets to illustrate the histopathological findings. In that same section, please add a sentence or two describing the lesion scoring concept/method. It is a nice added feature, not widespread in the field, and deserves a brief description.

      Edited as requested.  We added detailed description for scoring method in M&M under histopathology and lesion scoring

      Line 206: please add an introductory sentence explaining why one would expect S to cause (or not) hematological disruption, and why MCHC and RDW were chosen initially (they are markers of xyz). The first part of Figure 3 legend belongs to the Methods.

      To address this critique we added in #225-226 “The effect of L in the blood profile of humans and mouse has been reported (references #38-42 in revised manuscript) but the same has not been reported for S” . In line #229-230 we added “Of 20-blood parameters evaluated, two blood parameters were affected during treatment”. 

      The first part of Figure 3 legend belongs to the Methods.

      We edited Figure 3 to “During therapy of mice in Figure 1, the blood was collected at 1, 2- and 4-weeks posttreatment. The complete blood count was collected in VETSCAN® HM5 hematology analyzer (Zoetis)”.

      Line 218: please explain why the 4 blood parameters that are shown were selected, out of the 20 parameters surveyed.

      We added an explanation in line 239-240 “out 20-blood parameters evaluated, a total of four blood parameters were affected at 2 and 4-weeks-of treatment”.

      Line 243 and again Line 262 (similar to comment Line 206): please add an introductory paragraph explaining the motivation to conduct this analysis and the objective. Can the authors put the experiment in the context of their hypothesis?

      To address this critique, we added in line #235-237 “The Nix-TB trial associated the long-term administration of L within the BPaL regimen as the causative agent resulting in anemia in patients treated with the BPaL regimen (5).”

      Figure 4C (and the plasma and lung equivalent in the SI). This figure needs adequate labeling of axes: X axis = LOG CFU? Please add tick marks for all plots since log CFU is only shown for the bottom line. Y axes have no units: pg/mL as in B?

      Figure legend were edited to add (Y axis:pg/ml) and (X axis; log10CFU).  

      Line 255-256: please remove "pronounced" and "profound". There is a range of CFU reduction and cytokine reduction, from minor to major. The correlation trend is clear and those words are not needed.

      Edited as requested 

      Line 277-289, Figure 6: given the heterogeneity of a C3HeB/FeJ mouse lung (TB infected), and the very heterogeneous cell population distribution in these lungs (Fig. 6A), the validity of whole lung analysis on 2 or 3 mice (the legend should state what 1, 2 and 3 means, individual mice?) is put into question. "F4/80+ cells were observed significantly higher in BPaS compared to UnRx control": Figure S14 suggests a statistically significant difference, but nothing is said about the other cell type, which appears just as much reduced in BPaS compared to UnRx as F4/80+. Overall, sampling the whole lung for these analyses should be mentioned as a limitation in the Discussion.

      We agree with the reviewer that "visually" it appears as other populations in addition to F4/80 have statistical significance.  We run again the two way Anova with Tukey test and only the BPaS and UnRx for F4/80 is significant. 

      We edited figure S16 (previously S14) to add ns for every comparation.  

      In Figure 6A was edited ;  N=2 are 2 mice for Unrx and n=3 mice for BPaL/BPaS each.

      Line 355-360: "The BPa and BPaL regimens altered M:E in the C3HeB/FeJ TB model by suppressing myeloid and inducing erythroid lineages" This suggests that altered M:E is not associated with L, putting into question the comparison between BPaS, BPaL, and UnRx. Can the authors comment on how M:E is altered in BPa and not in BPaS?

      Our interpretation to this result was that addition of S in our regimen BPsS was capable of restoring the M:E ratio altered by the BPa and BPaL. This interpretation was included in main text in line #263-264 and is also now added to abstract

      Line 379: discuss the limitations of working with whole lungs.

      Sorry we cannot understand this request. In our studies we always work with whole lungs if the expected course of histopathology/infection among lung lobes is very variable (as is the case of C3HeB/Fej TB model)

      Concluding paragraph: "Here we present initial results that are in line with these goals." If such a bold claim is made, there needs to be a discussion on the translatability of the route of administration and the dose of S. Otherwise, please rephrase.

      We added the following last paragraph to discussion:

      To conclude, the TB drug development field is working towards developing shorter and safer therapies with a common goal of developing new multidrug regimens of low pill burden that are accessible to patients, of short duration (ideally 2-3 months) and consist of 3-4 drugs of novel mode-of-action with proven efficacy, safety, and limited toxicity. Here we present initial results for new multidrug regimens containing inhaled spectinamide 1599 that are in line with these goals. It is proposed that human use of spectinamides 1599 will be administered using a dry powder formulation delivered by the RS01 Plastiape dry powder inhaler.  We already reported on the aerodynamic properties of dry powder spectinamide 1599 within #3 HPMC capsules and delivered from a RS01 Plastiape inhaler device (reference #59 in revised manuscript). Future studies are already under consideration for funding by NIHNIAID to understand the pharmacokinetics of mono, binary and ternary combinations of BPaS. These studies also aim to identify the optimal dose level and dosing frequency of each regimen along with their efficacy and relapse free-sterilization potential. Studies are also planned using a model-based pharmacokinetic-pharmacodynamic (PKPD) framework, guided by an existing human BPa PKPD model (references #60 and 61 in revised manuscript) , to find allometric human dose levels, dosing frequencies and treatment durations that will inform the experimental design of future clinical studies.

      Minor edits

      Adverse events, not adverse effects (side effects)

      Edited as requested

      BALB/c (not Balb/c, please change throughout).

      Edited as requested

      Line 92: replace 'efficacy' with potency or activity.

      Edited as requested

      "Live" body weight: how is that different from "body weight"? Suggest deleting "live" throughout, or replace with "longitudinally recorded" if that's what is meant, although this is generally implied.

      Edited as requested

      The last line of Figure 2 legend is disconnected. 

      Line 331: delete "human".

      Edited as requested

      Reviewer #3 (Recommendations For The Authors):

      We thank the reviewer for these suggestions.  The data presented in this manuscript with 4 weeks of treatment along with monitoring of effects of therapy in blood, bone marrow and immunity have been submitted for a RO1 application to NIH-NIAID, which have received a fundable score and is under funding consideration. All the points suggested by the reviewer(s) here are included in the research proposed in the RO1 application including manufacturing and physico-chemically characterize larger scale of dry powders of spectinmides and evaluation of their aerodynamic performance for human or animal use; Pharmacokinetics and efficacy studies to determine the optimal dose level and dosing frequency for new multidrug regimens containing spectinamides. These studies include mono, binary and ternary combinations of each multidrug regimen along with their efficacy and relapse free- sterilization potential. These studies will also develop PK/PD simulation-based allometric scaling to aid in human dose projections inhalation. We hope the reviewer will understand all together these studies will last 4-5 years.  

      Although I truly appreciate the great efforts of the authors, I suggest that in order to better evaluate the contribution of S versus L to BPa in these models, repeat studies be run that:

      (a) include BPa groups to allow the contribution of S and L to be assessed. Included in research proposed RO1 application mentioned above

      (b) use shorter treatment times in BALB/c to allow comparisons at end of Tx CFU above the LOD. We have added new data for 2 weeks treatment with BPaL and BPaS in Balb/c mice infected with MTb that was removed from previous submission of this manuscript

      (c) use longer treatment times and ideally a relapse endpoint in Kramnik to allow

      assessment of L and S as contributors to BPa (i.e. give a chance to see better efficacy of BPaL or BPaS versus BPa) and also measure plasma exposures of all drugs (or lung levels if this is the translatable parameter for S) to allow detection of any large DDI and also understand the translation to the clinic. Related to the safety parameters, it would be really great to understand whether or not the observations for BPaL would be labeled adverse in a toxicology study/in a clinical study, and it would be useful to include information on the magnitude of observations seen here versus in the clinic (eg for the hematological parameters).

      The research proposed in the RO1 application mentioned above included extensive PK, extended periods of treatment beyond 1 month of treatment (2-5 months as needed to reach negative culturable bacterial from organs) and of course relapse studies. 

      Minor point: I suggest rewording "high safety profile" when describing spectinomides in the intro - or perhaps qualify the length of dosing where the drug is well tolerated

      "high safety profile" was replaced by “an acceptable safety profile”

    1. Author response:

      The following is the authors’ response to the original reviews.

      eLife assessment  

      This important study builds on a previous publication (with partially overlapping authors), demonstrating that T. brucei has a continuous endomembrane system, which probably facilitates high rates of endocytosis. Using a range of cutting-edge approaches, the authors present compelling evidence that an actomyosin system, with the myosin TbMyo1 as the molecular motor, is localized close to the endosomal system in the bloodstream form (BSF) of Trypanosoma brucei. It shows convincingly that actin is important for the organization and integrity of the endosomal system, and that the trypanosome Myo1is an active motor that interacts with actin and transiently associates with endosomes, but a role of Myo1 in endomembrane function in vivo was not directly demonstrated. This work should be of interest to cell biologists and microbiologists working on the cytoskeleton, and unicellular eukaryotes.

      We were delighted at the editors’ positive assessment and the reviewers’ rigorous, courteous, and constructive responses to the paper. We agree that a direct functional role for TbMyo1 in endomembrane activity was not demonstrated in the original submission, but have incorporated some new data (see new supplemental Figure S5) using the TbMyo1 RNAi cell line which are consistent with our earlier observations and interpretations.  

      Public Reviews:   

      Reviewer #1 (Public Review):  

      Using a combination of cutting-edge high-resolution approaches (expansion microscopy, SIM, and CLEM) and biochemical approaches (in vitro translocation of actin filaments, cargo uptake assays, and drug treatment), the authors revisit previous results about TbMyo1 and TbACT in the bloodstream form (BSF) of Trypanosoma brucei. They show that a great part of the myosin motor is cytoplasmic but the fraction associated with organelles is in proximity to the endosomal system. In addition, they show that TbMyo1 can move actin filaments in vitro and visualize for the first time this actomyosin system using specific antibodies, a "classical" antibody for TbMyo1, and a chromobody for actin. Finally, using latrunculin A, which sequesters G-actin and prevents F-actin assembly, the authors show the delocalization and eventually the loss of the filamentous actin signal as well as the concomitant loss of the endosomal system integrity. However, they do not assess the localization of TbMyo1 in the same conditions.  

      Overall the work is well conducted and convincing. The conclusions are not over-interpreted and are supported by the experimental results. 

      We are very grateful to Reviewer1 for their balanced assessment. The reviewer is correct that we did not assess the localisation of TbMyo1 following latrunculin A treatment, but it is worth noting that Spitznagel et al. carried out this exact experiment in the earlier 2010 paper – we have mentioned this in the revised manuscript.  

      Reviewer #2 (Public Review):  

      Summary:  

      The study by Link et al. advances our understanding of the actomyosin system in T. brucei, focusing on the role 

      of TbMyo1, a class I myosin, within the parasite's endosomal system. Using a combination of biochemical fractionation, in vitro motility assays, and advanced imaging techniques such as correlative light and electron microscopy (CLEM), this paper demonstrates that TbMyo1 is dynamically distributed across early and late endosomes, the cytosol, is associated with the cytoskeleton, and a fraction has an unexpected association with glycosomes. Notably, the study shows that TbMyo1 can translocate actin filaments at velocities suggesting an active role in intracellular trafficking, potentially higher than those observed for similar myosins in other cell types. This work not only elucidates the spatial dynamics of TbMyo1 within T. brucei but also suggests its broader involvement in maintaining the complex architecture of the endosomal network, underscoring the critical role of the actomyosin system in a parasite that relies on high rates of endocytosis for immune evasion. 

      Strengths:  

      A key strength of the study is its exceptional rigor and successful integration of a wide array of sophisticated techniques, such as in vitro motility assays, and advanced imaging methods, including correlative light and electron microscopy (CLEM) and immuno-electron microscopy. This combination of approaches underscores the study's comprehensive approach to examining the ultrastructural organization of the trypanosome endomembrane system. The application of functional data using inhibitors, such as latrunculin A for actin depolymerization, further strengthens the study by providing insights into the dynamics and regulatory mechanisms of the endomembrane system. This demonstrates how the actomyosin system contributes to cellular morphology and trafficking processes. Furthermore, the discovery of TbMyo1 localization to glycosomes introduces a novel aspect to the potential roles of myosin I proteins within the cell, particularly in the context of organelles analogous to peroxisomes. This observation not only broadens our understanding of myosin I functionality but also opens up new avenues for research into the cellular biology of trypanosomatids, marking a significant contribution to the field. 

      We are very pleased that the Reviewer felt the work is a significant contribution to the state of the art.  

      Weaknesses:  

      Certain limitations inherent in the study's design and scope render the narrative incomplete and make it challenging to reach definitive conclusions. One significant limitation is the reliance on spatial association data, such as colocalization of TbMyo1 with various cellular components-or the absence thereof-to infer functional relationships. Although these data suggest potential interactions, the authors do not confirm functional or direct physical interactions.  

      While TbMyo1's localization is informative, the authors do not directly demonstrate its biochemical or mechanical activities in vivo, leaving its precise role in cellular processes speculative. Direct assays that manipulate TbMyo1 levels, activity, and/or function, coupled with observations of the outcomes on cellular processes, would provide more definitive evidence of the protein's specific roles in T. brucei. A multifaceted approach, including genetic manipulations, uptake assays, kinetic trafficking experiments, and imaging, would offer a more robust framework for understanding TbMyo1's roles. This comprehensive approach would elucidate not just the "what" and "where" of TbMyo1's function but also the "how" and "why," thereby deepening our mechanistic insights into T. brucei's biology.  

      The reviewer is absolutely correct that the study lacks data on direct or indirect interactions between TbMyo1 and its intracellular partners, and this is an obvious area for future investigation. Given the generally low affinities of motor-cargo interactions, a proximity labelling approach (such has already been successfully used in studies of other myosins) would probably be the best way to proceed. 

      The reviewer is also right to highlight that a detailed mechanistic understanding of TbMyo1 function in vivo is currently lacking. We feel that this would be beyond the scope of the present work, but have included some new data using the TbMyo1 RNAi cell line (Figure S5), which are consistent with our previous findings.  

      Reviewer #3 (Public Review):  

      Summary:  

      In this work, Link and colleagues have investigated the localization and function of the actomyosin system in the parasite Trypanosoma brucei, which represents a highly divergent and streamlined version of this important cytoskeletal pathway. Using a variety of cutting-edge methods, the authors have shown that the T. brucei Myo1 homolog is a dynamic motor that can translocate actin, suggesting that it may not function as a more passive crosslinker. Using expansion microscopy, iEM, and CLEM, the authors show that MyoI localizes to the endosomal pathway, specifically the portion tasked with internalizing and targeting cargo for degradation, not the recycling endosomes. The glycosomes also appear to be associated with MyoI, which was previously not known. An actin chromobody was employed to determine the localization of filamentous actin in cells, which was correlated with the localization of Myo1. Interestingly, the pool of actomyosin was not always closely associated with the flagellar pocket region, suggesting that portions of the endolysomal system may remain at a distance from the sole site of parasite endocytosis. Lastly, the authors used actin-perturbing drugs to show that disrupting actin causes a collapse of the endosomal system in T. brucei, which they have shown recently does not comprise distinct compartments but instead a single continuous membrane system with subdomains containing distinct Rab markers.  

      Strengths:  

      Overall, the quality of the work is extremely high. It contains a wide variety of methods, including biochemistry, biophysics, and advanced microscopy that are all well-deployed to answer the central question. The data is also well-quantitated to provide additional rigor to the results. The main premise, that actomyosin is essential for the overall structure of the T. brucei endocytic system, is well supported and is of general interest, considering how uniquely configured this pathway is in this divergent eukaryote and how important it is to the elevated rates of endocytosis that are necessary for this parasite to inhabit its host.  

      We are very pleased that the Reviewer formed such a positive impression of the work. 

      Weaknesses:  

      (1) Did the authors observe any negative effects on parasite growth or phenotypes like BigEye upon expression of the actin chromobody?  

      Excellent question! There did appear to be detrimental effects on cell morphology in some cells, and it would definitely be worth doing a time course of induction to determine how quickly chromobody levels reach their maximum. The overnight inductions used here are almost certainly excessive, and shorter induction times would be expected to minimise any detrimental effects. We have noted these points in the Discussion.  

      (2) The Garcia-Salcedo EMBO paper cited included the production of anti-actin polyclonal antibodies that appeared to work quite well. The localization pattern produced by the anti-actin polyclonals looks similar to the chromobody, with perhaps a slightly larger labeling profile that could be due to differences in imaging conditions. I feel that the anti-actin antibody labeling should be expressly mentioned in this manuscript, and perhaps could reflect differences in the F-actin vs total actin pool within cells.  

      Implemented. We have explicitly mentioned the use of the anti-actin antibody in the Garcia-Salcedo paper in the revised Results and Discussion sections.  

      (3) The authors showed that disruption of F-actin with LatA leads to disruption of the endomembrane system, which suggests that the unique configuration of this compartment in T. brucei relies on actin dynamics. What happens under conditions where endocytosis and endocyctic traffic is blocked, such as 4 C? Are there changes to the localization of the actomyosin components? 

      Another excellent question! We did not analyse the localisation of TbMyo1 and actin under temperature block conditions, but this would definitely be a key experiment to do in follow-up work.

      (4) Along these lines, the authors suggest that their LatA treatments were able to disrupt the endosomal pathway without disrupting clathrin-mediated endocytosis at the flagellar pocket. Do they believe that actin is dispensable in this process? That seems like an important point that should be stated clearly or put in greater context.  

      Whether actin plays a direct or indirect role in endocytosis would be another fascinating question for future enquiry, and we do not have the data to do more than speculate on this point. Recent work in mammalian cells (Jin et al., 2022) has suggested that actin is primarily recruited when endocytosis stalls, and it could be that a similar role is at play here. We have noted this point in the Discussion. The observation of clathrin vesicles close to the flagellar pocket membrane and clathrin patches on the flagellar pocket membrane itself in the LatA-treated cells might suggest that some endocytic activity can occur in the absence of filamentous actin. 

      Recommendations for the authors:

      Note from the Reviewing Editor:  

      During discussion, all reviewers agreed that the role of TbMyo1 in vivo in endomembrane function had not been directly demonstrated. This could be done by testing the endocytic trafficking of (for example) fluorophoreconjugated TfR and BSA in the existing Myo1 RNAi line, using wide-field microscopy. Examining the endosomes/lysosomes' organization by thin-section EM would be even better. The actin signal detected by the chromobody tends to occupy a larger region than the MyoI. It's therefore conceivable that actin filamentation and stabilization via other actin-interacting proteins create the continuous endosomal structure, while MyoI is necessary for transport or other related processes. 

      These are all excellent points and very good suggestions. We have now incorporated new data (supplemental Figure S5) that includes BSA uptake assays in the TbMyo1 RNAi cell line and electron microscopy imaging after TbMyo1 depletion – the results are consistent with our earlier observations.   

      Reviewer #1 (Recommendations For The Authors):  

      -  Figure S2E. This panel is supposed to show the downregulation of TbMyo1 in the PCF compared to BSF but there is no loading control to support this claim. This is important because the authors mention in lines 381-383 that this finding conflicts with the previous study (Spitznagel et al., 2010). The authors also indicate in the figure legend that there is 50% less signal but there is no explanation about this quantification.   

      Good point. Equal numbers of cells were loaded in each lane, but we did not have an antibody against a protein known to be expressed at the same level in both PCF and BSF cells to use as a loading control. Using a total protein stain would have been similarly unhelpful in this context, as the proteomes of PCF and BSF cells are dissimilar. The quantification was made by direct measurement after background subtraction, but without normalisation owing to the lack of a loading control. This makes the conclusion somewhat tentative, but given the large difference in signal observed between the two samples (and the fact that this is consistent with the proteomic data obtained by Tinti and Ferguson) we feel that the conclusion is valid. We have clarified these points in the figure legend and Discussion.  

      -  It is mentioned in the discussion, as unpublished observations, that the predicted FYVE motif of TbMyo1 can bind specifically PI(3)P lipids. This is a very interesting point that would be new and would strengthen the suggested association with the endosomal system mainly based on imaging data. 

      We agree that this is – potentially – a very exciting observation and it is an obvious direction for future enquiry.  

      The data are preliminary at this stage and will form the basis of a future publication. Given that the predicted FYVE domain of TbMyo1 and known lipid-binding activity of other class I myosins makes this activity not wholly unexpected, we feel that it is acceptable at this stage to highlight these preliminary findings.  

      -  The authors use the correlation coefficient to estimate the colocalization (lines 223-226). Although they clearly explain the difference between the correlation coefficient and the co-occurrence of two signals, I wonder if it would not be clearer for the audience to have quantification of the overlapping signals. Also, it is not mentioned on which images the correlation coefficient was measured. It seems that it is from widefield images (Figures 3E and 6E), and likely from SIM images for Figure 3C but the resolution is different. Are widefield images sufficient to assess these measurements? 

      With hindsight, and given the different topological locations of TbMyo1 and the cargo proteins (cytosolic and lumenal, respectively) it would probably have been wiser to measure co-occurrence rather than correlation, but we would prefer not to repeat the entire analysis at this stage. The correlations were measured from widefield images using the procedure described in the Materials & Methods. These are obviously lower resolution than confocal or SIM images would be, but are still of value, we believe. One further point – upon re-examination of some of the TbMyo1 transferrin (Tf) and BSA data, we noticed that there are many pixels with a value of 0 for Tf/BSA and a nonzero value for TbMyo1 and vice-versa. The incidence of zero-versus-nonzero values in the two channels will have lowered the correlation coefficient, and in this sense, the correlation coefficients are giving us a hint of what the immuno-EM images later confirm: that the TbMyo1 and cargo are present in the same locations, but in different proportions. We have added this point to the discussion.  

      -  It would be good to know if the loss of the endosomal system integrity (using EBI) is the same upon TbMyo1 depletion than in the latrunculin A treated parasites. 

      We agree! We have now included new data (Figure S5) that suggests endosomal system morphology is altered upon TbMyo1 depletion. We would predict that the effect upon TbMyo1 depletion is slower or less dramatic than upon LatA treatment (as LatA affects both actin and TbMyo1, given that TbMyo1 depends upon actin for its localisation).

      -  Conversely, it would be of interest to see how the localization of TbMyo1 changes upon latrunculin A treatment.

      This experiment was done in 2010 by Spitznagel et al., who observed a delocalisation of the TbMyo1 signal after LatA treatment. We have noted this in the Results and Discussion.

      Minor corrections:  

      -  Line 374: Figure S1 should be Figure S2. 

      Implemented (many thanks!).  

      -  Panel E of Figure S2 refers to TbMyo1 and should therefore be included in Figure S1 and not S2. 

      We would prefer not to implement this suggestion. We did struggle over the placing of this panel for exactly this reason, but as the samples were obtained as part of the experiments described in Figure S2, we felt that its placement here worked best in terms of the narrative of the manuscript.    

      -  Figure S2F: the population of TbMyo21 +Tet seems lost after 48 h although the authors mention that there is no growth defect. 

      Good eyes! We have re-added the panel, which shows that there was no growth defect in the tetracycline-treated population.  

      Reviewer #2 (Recommendations For The Authors):  

      Fig 1 vs. Figure 3: The biochemical fractionation experiments have been well-controlled, showing that 40% of TbMyo1 is found in both the cytosolic and cytoskeletal fractions, with only 20% in the organelle-associated fraction. The conclusion is supported by the experimental design, which includes controls to rule out crosscontamination between fractions. However, does this contrast with the widefield microscopy experiments, where the vast majority of the signal is in endocytic compartments and nowhere else? 

      This is a good point. There are three factors that probably explain this. First, given that the actin cytoskeleton is associated with the endosomal system, a large proportion of the material partitioning into the cytoskeleton (P2) fraction is probably localised to the endosomal system (a fun experiment would be to repeat the fractionation with addition of ATP to the extraction buffer to make the myosin dissociate and see whether more appeared in the SN2 fraction as a result). Second, the 40% of the TbMyo1 that is cytosolic is distributed throughout the entire cellular volume, whereas the material localised to the endosomes is concentrated in a much smaller space, by comparison, and producing a stronger signal. Third, the widefield microscopy images have had brightness and contrast adjusted in order to reduce “background” signal, though this will also include cytosolic molecules. We hope these explanations are satisfactory, but would welcome any additional thoughts from either the reviewer or the community.  

      The section title 'TbMyo1 translocates filamentous actin at 130 nm/s' could mislead readers by not specifying that the findings are from an in vitro experiment with a recombinant protein, which may not fully reflect the cell's complex context. Although this detail is noted in the figure legend, incorporating it into the main text and considering a title revision would ensure clarity and accuracy.  

      Good point. Implemented – we have amended the section title to “TbMyo1 translocates filamentous actin at 130 nm/s in vitro” and the figure legend title to “TbMyo1 translocates filamentous actin in vitro”.  

      The discussion of the translocation experiment could be better phrased addressing certain limitations. The in vitro conditions might not fully capture the complexity and dynamic nature of cellular environments where multiple regulatory mechanisms, interacting partners, and cellular compartments come into play. 

      Good point, implemented. We have added a note on this to the Discussion.  

      It is puzzling that RNAi, which is widely used in T. brucei was not used to further investigate the functional roles of TbMyo1 in Trypanosoma brucei. Given that the authors already had the cell line and used it to validate the specificity of the anti-TbMyo1. RNAi could have been employed to knock down TbMyo1 expression and observe the resultant effects on actin filament dynamics and organization within the cell. This would have directly tested TbMyo1's contribution to actin translocation observed in the in vitro experiments. 

      It would obviously be interesting to carry out an in-depth characterisation of the phenotype following TbMyo1 depletion and whether this has an effect on actin dynamics. We have now included additional data (supplemental Figure S5) using the TbMyo1 RNAi cells and the results are consistent with our earlier observations and interpretations. It is worth noting too that at least for electron microscopy studies of intracellular morphology, the slower onset of an RNAi phenotype and the asynchronous replication of T. brucei populations make observation of direct (early) effects of depletion challenging – hence the preferential use of LatA here to depolymerise actin and trigger a faster phenotype.  

      I found that several declarative statements within the main text may not be fully supported by the overall evidence. I suggest modifications to present a more balanced view,  

      Line 227: "The results here suggest that although the TbMyo1 distribution overlaps with that of endocytic cargo, the signals are not strongly correlated." This conclusion about the lack of strong correlation might mislead readers about the functional relationship between TbMyo1 and endocytic cargo, as colocalization does not directly imply functional interaction. 

      We would prefer not to alter this statement. It was our intention to phrase this cautiously, as we have not directly investigated the functional interplay between TbMyo1 and endocytic cargo and the subsequent sentence directs the reader to the Discussion for more consideration of this issue.    

      Line 397: "This relatively high velocity might indicate that TbMyo1 is participating in intracellular trafficking of BSF T. brucei and functioning as an active motor rather than a static tether." The statement directly infers TbMyo1's functional role from in vitro motility assay velocities without in vivo corroboration.

      We have amended the sentence in the Discussion to make it clear that it is speculative.  

      The hypothesis that cytosolic TbMyo1 adopts an auto-inhibited "foldback" configuration, drawn by analogy with findings from other studies, is intriguing. Yet, direct evidence linking this configuration to TbMyo1's function in T. brucei is absent from the data presented. 

      We have amended the sentence in the Discussion to make it clear that it is speculative. Future in vitro experiments will test this hypothesis directly.  

      The suggestion that a large cytosolic fraction of TbMyo1 indicates dynamic behavior, high turnover on organelles, and a low duty ratio is plausible but remains speculative without direct experimental evidence. Measurements of TbMyo1 turnover rates or duty ratios in T. brucei through kinetic studies would substantiate this claim with the necessary evidence.  

      We have amended the sentence in the Discussion to make it clear that it is speculative, and deleted the reference to a possible low duty ratio. Again, future in vitro experiments will measure the duty ratio of TbMyo1 using stopped-flow. 

      Reviewer #3 (Recommendations For The Authors):  

      Lines 171-172: The authors mention that MyoI could be functioning as a motor rather than a tether. The differences in myosin function have not been introduced prior to this. I would recommend explaining these differences and what it could mean for the function of the motor in the introduction to help a non-expert audience.

      Good point. Implemented.  

      Line 94-95: This phenotype only holds for the bloodstream form- the procyclic form are quite resistant to actin RNAi and MyoI RNAi. I would clarify. 

      Good point. Implemented.  

      Line 142-146: did the authors attempt to knock out the Myo21? 

      Good point. No, this was not attempted. Given the extremely low expression levels of TbMyo21 in the BSF cells we would not expect a strong phenotype, but this assumption would be worth testing. 

      Figure 3D: is there a reason why the authors chose to show the single-channel images in monochrome in this case?  

      Not especially. These panels are the only ones that show a significant overlap in the signals between the two channels (unlike the colabelling experiments with ER, Golgi), so greyscale images were used because of their higher contrast. 

      Line 397-398: I'm struggling a bit to understand how MyoI could be involved in intracellular trafficking in the endosomal compartments if the idea is that we have a continuous membrane? Some more detail as to the author's thinking here would be useful. 

      Implemented. We have noted that this statement is speculative, and emphasised that being an active motor does not automatically mean that it is involved in intracellular traffic – it could instead be involved in manipulating endosomal membranes. We have noted too that the close proximity between TbMyo1 and the lysosome (Figures

      3-5) could be important in this regard. The lysosome is not contiguous with the endosomal system, and it is possible that TbMyo1 is working as a motor to transport material (class II clathrin-coated vesicles) from the endosomal system to the lysosome.  

      Line 493-496: Does this mean that endocytosis from the FP does not require actin? This would be hard to explain considering the phenotypes observed in the original actin RNAi work. Is the BigEye phentopye observed in BSF actin RNAi and Myo1 RNAi cells due to some indirect effect? 

      It seems possible that actin is not directly or essentially involved in endocytosis, and the characterisation of the actin RNAi phenotype would be worth revisiting in this respect – we have noted this in the Discussion. Although RNAi of actin was lethal, the phenotype appears less penetrant than that seen following depletion of the essential endocytic cofactor clathrin (based on the descriptions in Garcia-Salcedo et al., 2004 and Allen et al., 2003). BigEye phenotypes occur in BSF cells whenever there is some perturbation of endomembrane trafficking and are not necessarily a direct consequence of depletion – this is why careful investigation of early timepoints following RNAi induction is critical.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews: 

      Reviewer #1 (Public Review): 

      Summary: 

      This manuscript presented a useful toolkit designed for CyTOF data analysis, which integrates 5 key steps as an analytical framework. A semi-supervised clustering tool was developed, and its performance was tested in multiple independent datasets. The tool was compared to human experts as well as supervised and unsupervised methods. 

      Strengths: 

      The study employed multiple independent datasets to test the pipeline. A new semi-supervised clustering method was developed. 

      Weaknesses: 

      The examination of the whole pipeline is incomplete. Lack of descriptions or justifications for some analyses. 

      We thank the reviewer’s overall summary and comments of this manuscript. In the last part of the results, we showcased the functionalities of ImmCellTyper in covid dataset, including quality check, BinaryClust clustering, cell abundance quantification, state marker expression comparison within each identified cell types, cell population extraction, subpopulation discovery using unsupervised methods, and data visualization etc. We added more descriptions in the text based on the reviewer’s suggestions. 

      Reviewer #2 (Public Review): 

      Summary: 

      The authors have developed marker selection and k-means (k=2) based binary clustering algorithm for the first-level supervised clustering of the CyTOF dataset. They built a seamless pipeline that offers the multiple functionalities required for CyTOF data analysis. 

      Strengths: 

      The strength of the study is the potential use of the pipeline for the CyTOF community as a wrapper for multiple functions required for the analysis. The concept of the first line of binary clustering with known markers can be practically powerful. 

      Weaknesses: 

      The weakness of the study is that there's little conceptual novelty in the algorithms suggested from the study and the benchmarking is done in limited conditions. 

      We thank the reviewer’s overall summary and comments of this manuscript. While the concept of binary clustering by k-means is not novel, BinaryClust only uses it for individual markers to identify positive and negative cells, then combine it with the pre-defined matrix for cell type identification. This has not been introduced elsewhere. Furthermore, ImmCellTyper streamlines the entire analysis process and enhances data exploration on multiple levels. For instance, users can evaluate functional marker expression level/cellular abundance across both main cell types and subpopulations; Also, this computational framework leverages the advantages of both semi-supervised and unsupervised clustering methods to facilitate subpopulation discovery. We believe these contributions warrant consideration as advancements in the field.  

      As for the benchmarking, we limited the depth only to main cell types rather than subpopulations. The reason is because we only apply BinaryClust to identify main cell types; For the cell subsets discovery, unsupervised methods integrated in this pipeline has already been published and widely used by the research community. Therefore, it does not seem to be necessary for additional benchmarking.

      Reviewer #3 (Public Review): 

      Summary: 

      ImmCellTyper is a new toolkit for Cytometry by time-of-flight data analysis. It includes BinaryClust, a semi-supervised clustering tool (which takes into account prior biological knowledge), designed for automated classification and annotation of specific cell types and subpopulations. ImmCellTyper also integrates a variety of tools to perform data quality analysis, batch effect correction, dimension reduction, unsupervised clustering, and differential analysis. 

      Strengths: 

      The proposed algorithm takes into account the prior knowledge. 

      The results on different benchmarks indicate competitive or better performance (in terms of accuracy and speed) depending on the method. 

      Weaknesses: 

      The proposed algorithm considers only CyTOF markers with binary distribution. 

      We thank the reviewer’s overall summary and comments of this manuscript. Binary classification can be considered as an imitation of human gating strategy, as it is applied to each marker. For example, when characterizing the CD8 T cells, we aim for CD19-CD14-CD3+CD4- population, which is binary in nature (either positive and negative) and follows the same logic as the method (BinaryClust) we developed. Results indicated that it works very well for well-defined main cell lineages, particularly when the expression of the defining marker is not continuous. However, the limitation is for subpopulation identification, because a handful makers behave in a continuum manner, so we suggest unsupervised method after BinaryClust, which also brings another advantage of identifying unknown subsets beyond our current knowledge, and none of the semi-supervised tools can achieve that. To address the reviewer’s concern, we considered the limitation of binary distribution, but it does not profoundly affect the application of the pipeline.

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors): 

      Many thanks for the reviewers’ comments and suggestions, please see below the point-to-point response:

      (1) The style of in-text reference citation is not consistent. Many do not have published years.

      The style of the reference citation has been revised and improved.  

      (2) The font size in the table of Figure 1 is too small, so is Figure 2. 

      The font size has been increased.

      (3) Is flowSOM used as part of BinaryClust? How should the variable running speed of BinaryClust be interpreted, given that it is occasionally slower and sometimes faster than flowSOM in the datasets?

      To answer reviewer’s question, flowSOM is not a part of BinaryClust. They are separate clustering methods that have been incorporated into the ImmCellTyper pipeline. As described in Figure 1, BinaryClust, a semi-supervised method, is used to classify the main cell lineages; while flowSOM, an unsupervised method, is recommended here for further subpopulation discovery. So, they operate independently of each other. To avoid confusions, we slightly modified Figure 1 for clarification.

      Regarding the variability in running speed in Figure 4. The performance of algorithms can indeed be influenced by the characteristics of the datasets, such as size and complexity. The differences observed between the covid dataset and the MPN dataset, such as marker panel, experimental protocol, and data acquisition process etc., could account for this variation. Our explanation is that flowSOM suits better the data structure of covid dataset, which might be the reason why it is slightly faster to analyse compared to the MPN dataset. Moreover, for the covid dataset, the runtime for both BinaryClust and flowSOM is less than 100s, and the difference is not notable. 

      (4) In the Method section ImmCellTyper workflow overview, it is difficult to link the description of the pipeline to Figure 8. There are two sub-pipelines in the text and seven steps in the figure. What are their relations? Some steps are not introduced in the text, such as Data transformation and SCE object construction. What is co-factor 5?

      Figure 8 provides an overview of the entire workflow for CyTOF data analysis, starting from the raw fcs file data and proceeding until downstream analysis (seven steps). But the actual implementation of the pipeline was divided into two separate sections, as outlined in the vignettes of the ImmCellTyper GitHub page (https://github.com/JingAnyaSun/ImmCellTyper/tree/main/vignettes).

      Users will initially run ‘Intro_to_batch_exam_correct’ to perform data quality check and identify potential batch effects, followed by ‘Intro_to_data_analysis’ for data exploration. We agree with the reviewer that the method for this section is a bit confusing, so we’ve added more description for clarification.

      In processing mass cytometry data, arcsine transformation is commonly applied to handle zero values, skewed distributions, and to improve visualization as well as clustering performance. The co-factor here is used as a parameter to scale down the data to control the width of the linear region before arcsine transformation. We usually get the best results by using co-factor 5 for CyTOF data.   

      (5) For differential analysis, could the pipeline analyze paired/repeated samples?

      For the statistical step, ImmCellTyper supports both two-study group comparison using Mann-Whitney Wilcoxon test, and multiple study group comparison (n>2) using Kruskal Wallis test followed by post hoc analysis (pairwise Wilcoxon test or Dunn’s test) with multiple testing correction using Benjamini-Hochberg Procedure.

      Certainly, this pipeline allows flexibilities, users can also extract the raw data of cell frequencies and apply suitable statistical methods for testing.

      (6) In Figure 2A, the range of the two axes is different for Dendritic cells, which could be misleading. Why the agreement is bad for dendritic cells?

      The range for the axes is automatically adapted to the data structure, which explains why they may not necessarily be equal. The co-efficient factor for the correlation of DCs is 0.958, compared to other cell types (> 0.99), it is relatively worse but does not indicate poor agreement.

      Moreover, the abundance of DCs is much less than other cell types, comprising approximately 2-5% of whole cells. As a result, even small differences in abundance may appear to as significant variations. For example, a difference of 1% in DC abundance represents a 2-fold change, which can be perceived as substantial.

      Overall, while the agreement for DCs may appear comparatively lower, it is not necessarily indicative of poor performance, considering both the coefficient factor and the relative abundance of DCs compared to other cell types.

      (7) In the Results section BinaryClust achieves high accuracy, what method was used to get the p-value, such as lines 212, 213, etc.?

      The accuracy of BinaryClust was tested using F-measure and ARI against ground truth (manual gating), the detailed description/calculation can be found in methods. For line 212 and 213, the p-value was calculated using ANOVA for the interaction plot shown in Figure 3. We’ve now added the statistical information into the figure legend.   

      (8) The performance comparison between BinaryClust and LDA is close. The current comparison design looks unfair. Given LDA only trained using half data, LDA may outperform BinaryClust.

      It is true that LDA was trained using half data, which is because this method requires manual gating results as training dataset to build a model, then apply the model to the rest of the files to label cell types. Here we used 50% of the whole dataset as training set. We are of course very happy to implement any additional suggestions for a better partition ratio.

      (9) There are 5 key steps in the proposed workflow. However, not every step was presented in the Results.

      Thanks for the comments. The results primarily focused on demonstrating the precision and performance of BinaryClust in comparison with ground truth and existing tools. Additionally, a case study showcasing the application/functions of the entire pipeline in a dataset was also presented. Due to limitation in space, the implementation details of the pipeline were described in the method section and github documentations, which users/readers can easily access.

      Reviewer #2 (Recommendations For The Authors): 

      The tools suggested by the authors could be potentially useful to the community. However, it's difficult to understand the conceptual novelty of the algorithms suggested here. The concept of binary clustering has been described before (https://doi.org/10.1186/s12859-022-05085-zhttps://doi.org/10.1152/ajplung.00104.2022), and it mainly utilizes k-means clustering set to generate binary clusters based on selected markers. Other algorithms associated with the package are taken from other studies. 

      We acknowledge the reviewer’s comment regarding the novelty of our method. While the concept of binary clustering by k-means has been previously described to transcriptome data, our approach applies it to CyTOF data analysis, which has not been introduced elsewhere. Furthermore, ImmCellTyper streamlines the entire analysis process and enhances data exploration on multiple levels. For instance, users can evaluate functional marker expression level/cellular abundance across both main cell types and subpopulations; Also, as stated in the manuscript, this computational framework leverages the advantages of both semi-supervised and unsupervised clustering methods to facilitate subpopulation discovery. We believe these contributions warrant consideration as advancements in the field.  

      In addition, the benchmarking of clustering performance, especially to reproduce manual gating and comparison to tools such as flowSOM is not comprehensive enough. The result for the benchmarking test could significantly vary depending on how the authors set the ground truth (resolution of cell type annotations). The authors should compare the tool's performance by changing the depth of cell type annotations. Especially, the low abundance cell types such as gdT cells or DCs were not effectively captured by the suggested methods. 

      Thanks for the comment. We appreciate the reviewer’s concern. However, as illustrated in figure 1, our approach uses BinaryClust, a semi-supervised method, to identify main cell types rather than directly targeting subpopulations. The reason is because semi-supervised method relies on users’ prior definition thus is limited to discover novel subsets. In the ImmCellTyper framework, unsupervised method was subsequently applied for subset exploration following the BinaryClust step.

      Regarding benchmarking, we focused on testing the precision of BinaryClust for main cell type characterization, because it is what the method is used for in the pipeline, and we believe this is sufficient. As for the cell subsets discovery, the unsupervised methods we integrated has already been published and widely used by the research community. Therefore, it does not seem to be necessary for additional benchmarking.

      Moreover, as shown in Figure 3 and Table 1, our results indicated that the F-measure for DCs and gdT cells in BinaryClust is 0.80 and 0.92 respectively, which were very close to ground truth and outperformed flowSOM, demonstrating its effectiveness. 

      We hope these clarifications address the reviewer’s concern.

      Minor comments: 

      (1) In Figure 4, it's perplexing to note that BinaryClust shows the slowest runtime for the COVID dataset, compared to the MPN dataset, which features a similar number of cells. What causes this variation? Is it dependent on the number of markers utilized for the clustering? This should be clarified/tested. 

      Thanks for the comment, but we are not sure that we fully understand the question. As shown in figure 4 that BinaryClust has slightly higher runtime in MPN dataset than covid dataset, which is reasonable because and the cell number in MPN dataset is around 1.6 million more than covid dataset.

      (2) Some typos are noted: 

      - DeepCyTOF and LDA use a maker expression matrix extracted → "marker"?* 

      Corrected.

      - Datasets(Chevrier et al.)which → spacing* 

      Corrected.

      - This is due to the method's reliance → spacing*

      Corrected.

      Reviewer #3 (Recommendations For The Authors): 

      Is it possible to accommodate more than two levels within the clustering process, i.e., can the proposed semi-supervised clustering tool be extended to multi-levels instead of binary?

      Thanks for the comments. Binary classification can be considered as an imitation of human gating strategy, as it is applied to each marker. For example, when characterizing the CD8 T cells, we aim for CD19-CD14-CD3+CD4- population, which is binary in nature (either positive and negative) and follows the same logic as the method (BinaryClust) we developed. Results indicated that it works very well for well-defined main cell lineages. However, the limitation is for subpopulation identification, because a handful of makers behave in a continuum manner, so we would suggest unsupervised method after BinaryClust, which also brings another advantage of identifying unknown subsets beyond our current knowledge, and none of the semi-supervised tools can achieve that. To answer the reviewer’s question, it is possible to set the number to 3,4,5 rather than just 2, but considering the design and rationale of the entire framework (as describe in the manuscript and above), it doesn’t seem to be necessary.

      Could you please comment on why on the COVID dataset, BinaryClust was slower as compared to flowSOM?

      Thanks for the question. The performance of algorithms can indeed be affected by the characteristics of the datasets, such as their size and complexity. The covid and MPN datasets differ in various aspects including marker panel, experimental protocol, and data acquisition process, among others, which wound account for the observed variation in speed. So, our explanation is flowSOM suits better for the structure of covid dataset than MPN dataset.  Additionally, for covid dataset, both BinaryClust and flowSOM have runtimes of less than 100s, and the difference between the two isn’t particularly dramatic.

      Minor errors: 

      Line#215 "(ref) " reference is missing

      Added.

      Figure 3, increase the font of the text in order to improve readability. 

      Increased.

      Line#229 didn't --> did not. 

      Corrected

      Line#293 repetition of the reference. 

      The repetition is due to the format of the citation, which has been revised.

    1. Author response:

      The following is the authors’ response to the previous reviews.

      Reviewer 1:

      Thank you for your review and pointing out multiple things to be discussed and clarified! Below, we go through the various limitations you pointed out and refer to the places where we have tried to address them.

      (1) It's important to keep in mind that this work involves simplified models of the motor system, and often the terminology for 'motor cortex' and 'models of motor cortex' are used interchangeably, which may mislead some readers. Similarly, the introduction fails in many cases to state what model system is being discussed (e.g. line 14, line 29, line 31), even though these span humans, monkeys, mice, and simulations, which all differ in crucial ways that cannot always be lumped together.

      That is a good point. We have clarified this in the text (Introduction and Discussion), to highlight the fact that our model isn’t necessarily meant to just capture M1. We have also updated the introduction to make it more clear which species the experiments which motivate our investigation were performed in.

      (2) At multiple points in the manuscript thalamic inputs during movement (in mice) is used as a motivation for examining the role of preparation. However, there are other more salient motivations, such as delayed sensory feedback from the limb and vision arriving in the motor cortex, as well as ongoing control signals from other areas such as the premotor cortex.

      Yes – the motivation for thalamic inputs came from the fact that those have specifically been shown to be necessary for accurate movement generation in mice. However, it is true that the inputs in our model are meant to capture any signals external to the dynamical system modeled, and as such are likely to represent a mixture of sensory signals, and feedback from other areas. We have clarified this in the Discussion, and have added this additional motivation in the Introduction.

      (3) Describing the main task in this work as a delayed reaching task is not justified without caveats (by the authors' own admission: line 687), since each network is optimized with a fixed delay period length. Although this is mentioned to the reader, it's not clear enough that the dynamics observed during the delay period will not resemble those in the motor cortex for typical delayed reaching tasks.

      Yes, we completely agree that the terminology might be confusing. While the task we are modeling is a delayed reaching task, it does differ from the usual setting since the network has knowledge of the delay period, and that is indeed a caveat of the model. We have added a brief paragraph just after the description of the optimal control objective to highlight this limitation.

      We have also performed additional simulations using two different variants of a model-predictive control approach that allow us to relax the assumption that the go-cue time is known in advance. We show that these modifications of the optimal controller yield results that remain consistent with our main conclusions, and can in fact in some settings lead to preparatory activity plateaus during the preparation epoch as often found in monkey M1 (e.g in Elsayed et al. 2016). We have modified the Discussion to explain these results and their limitations, which are summarized in a new Supplementary Figure (S9).

      (4) A number of simplifications in the model may have crucial consequences for interpretation.

      a) Even following the toy examples in Figure 4, all the models in Figure 5 are linear, which may limit the generalisability of the findings.

      While we agree that linear models may be too simplistic, much prior analyses of M1 data suggest that it is often good enough to capture key aspects of M1 dynamics; for example, the generative model underlying jPCA is linear, and Sussillo et al. (2015) showed that the internal activity of nonlinear RNN models trained to reproduce EMG data aligned best with M1 activity when heavily regularized; in this regime, the RNN dynamics were close to linear. Nevertheless, this linearity assumption is indeed convenient from a modeling viewpoint: the optimal control problem is more easily solved for linear network dynamics and the optimal trajectories are more consistent across networks. Indeed, we had originally attempted to perform the analyses of Figure 5 in the nonlinear setting, but found that while the results were overall similar to what we report in the linear regime, iLQR was occasionally trapped into local minimal, resulting in more variable results especially for inhibition-stabilized network in the strongly connected end of the spectrum. Finally, Figure 5 is primarily meant to explore to what extent motor preparation can be predicted from basic linear control-theoretic properties of the Jacobian of the dynamics; in this regard, it made sense to work with linear RNNs (for which the Jacobian is constant).

      b) Crucially, there is no delayed sensory feedback in the model from the plant. Although this simplification is in some ways a strength, this decision allows networks to avoid having to deal with delayed feedback, which is a known component of closed-loop motor control and of motor cortex inputs and will have a large impact on the control policy.

      This comment resonates well with Reviewer 3's remark regarding the autonomous nature (or not) of M1 during movement. Rather than thinking of our RNN models as anatomically confined models of M1 alone, we think of them as models of the dynamics which M1 implements possibly as part of a broader network involving “inter-area loops and (at some latency) sensory feedback”, and whose state appears to be near-fully decodable from M1 activity alone. We have added a paragraph of Discussion on this important point.

      (5) A key feature determining the usefulness of preparation is the direction of the readout dimension. However, all readouts had a similar structure (random Gaussian initialization). Therefore, it would be useful to have more discussion regarding how the structure of the output connectivity would affect preparation, since the motor cortex certainly does not follow this output scheme.

      We agree with this limitation of our model — indeed one key message of Figure 4 is that the degree of reliance on preparatory inputs depends strongly on how the dynamics align with the readout. However, this strong dependence is somewhat specific to low-dimensional models; in higher-dimensional models (most of our paper), one expects that any random readout matrix C will pick out activity dimensions in the RNN that are sufficiently aligned with the most controllable directions of the dynamics to encourage preparation.

      We did consider optimizing C away (which required differentiating through the iLQR optimizer, which is possible but very costly), but the question inevitably arises what exactly should C be optimized for, and under what constraints (e.g fixed norm or not). One possibility is to optimize C with respect to the same control objective that the control inputs are optimized for, and constrain its norm (otherwise, inputs to the M1 model, and its internal activity, could become arbitrarily small as C can grow to compensate). We performed this experiment (new Supplementary Figure S7) and obtained a similar preparation index; there was one notable difference, namely that the optimized readout modes led to greater observability compared to a random readout; thus, the same amount of “muscle energy” required for a given movement could now be produced by a smaller initial condition. In turn, this led to smaller control inputs, consistent with a lower control cost overall.

      Whilst we could have systematically optimized C away, we reasoned that (i) it is computationally expensive, and (ii) the way M1 affects downstream effectors is presumably “optimized” for much richer motor tasks than simple 2D reaching, such that optimizing C for a fixed set of simple reaches could lead to misleading conclusions. We therefore decided to stick with random readouts.

      Additional comments:

      (1) The choice of cost function seems very important. Is it? For example, penalising the square of u(t) may produce very different results than penalising the absolute value.

      Yes, the choice of cost function does affect the results, at least qualitatively. The absolute value of the inputs is a challenging cost to use, as iLQR relies on a local quadratic approximation of the cost function. However, we have included additional experiments in which we penalized the squared derivative of the inputs (Supplementary Figure S8; see also our response to Reviewer 3's suggestion on this topic), and we do see differences in the qualitative behavior of the model (though the main takeaway, i.e. the reliance on preparation, continues to hold). This is now referred to and discussed in the Discussion section.

      (2) In future work it would be useful to consider the role of spinal networks, which are known to contribute to preparation in some cases (e.g. Prut and Fetz, 1999).

      (3) The control signal magnitude is penalised, but not the output torque magnitude, which highlights the fact that control in the model is quite different from muscle control, where co-contraction would be a possibility and therefore a penalty of muscle activation would be necessary. Future work should consider the role of these differences in control policy.

      Thank you for pointing us to this reference! Regarding both of these concerns, we agree that the model could be greatly improved and made more realistic in future work (another avenue for this would be to consider a more realistic biophysical model, e.g. using the MotorNet library). We hope that the current Discussion, which highlights the various limitations of our modeling choices, makes it clear that a lot of these choices could easily be modified depending on the specific assumptions/investigation being performed.

      Reviewer 2:

      Thank you for your positive review! We very much agree with the limitations you pointed out, some of which overlapped with the comments of the other reviewers. We have done our best to address them through additional discussion and new supplementary figures. We briefly highlight below where those changes can be found.

      (1) Though the optimal control theory framework is ideal to determine inputs that minimize output error while regularizing the input norm, it however cannot easily account for some other varied types of objectives especially those that may lead to a complex optimization landscape. For instance, the reusability of parts of the circuit, sparse use of additional neurons when learning many movements, and ease of planning (especially under uncertainty about when to start the movement), may be alternative or additional reasons that could help explain the preparatory activity observed in the brain. It is interesting to note that inputs that optimize the objective chosen by the authors arguably lead to a trade-off in terms of other desirable objectives. Specifically, the inputs the authors derive are time-dependent, so a recurrent network would be needed to produce them and it may not be easy to interpolate between them to drive new movement variants. In addition, these inputs depend on the desired time of output and therefore make it difficult to plan, e.g. in circumstances when timing should be decided depending on sensory signals. Finally, these inputs are specific to the full movement chain that will unfold, so they do not permit reuse of the inputs e.g. in movement sequences of different orders.

      Yes, that is a good point! We have incorporated further Discussion related to this point. We have additionally included a new example in which we regularize the temporal complexity of the inputs (see also our response to Reviewer 3's suggestion on this topic), which leads to more slowly varying inputs, and may indeed represent a more realistic constraint and lead to simpler inputs that can more easily be interpolated between. We also agree that uncertainty about the upcoming go cue may play an important role in the strategy adopted by the animals. While we have not performed an extensive investigation of the topic, we have included a Supplementary Figure (S9) in which we used Model Predictive Control to investigate the effect of planning under uncertainty about the go cue arrival time. We hope that this will give the reader a better sense of what sort of model extensions are possible within our framework.

      (2) Relatedly, if the motor circuits were to balance different types of objectives, the activity and inputs occurring before each movement may be broken down into different categories that may each specialize into one objective. For instance, previous work (Kaufman et al. eNeuron 2016, Iganaki et al., Cell 2022, Zimnik and Churchland, Nature Neuroscience 2021) has suggested that inputs occurring before the movement could be broken down into preparatory inputs 'stricto sensu' - relating to the planned characteristics of the movement - and a trigger signal, relating to the transition from planning to execution - irrespective of whether the movement is internally timed or triggered by an external event. The current work does not address which type(s) of early input may be labeled as 'preparatory' or may be thought of as a part of 'planning' computations.

      Yes, our model does indeed treat inputs in a very general way, and does not distinguish between the different types of processes they may be composed of. This is partly because we do not explicitly model where the inputs come from, such that our inputs likely englobe multiple processes. We have added discussion related to this point.

      (3) While the authors rightly point out some similarities between the inputs that they derive and observed preparatory activity in the brain, notably during motor sequences, there are also some differences. For instance, while both the derived inputs and the data show two peaks during sequences, the data reproduced from Zimnik and Churchland show preparatory inputs that have a very asymmetric shape that really plummets before the start of the next movement, whereas the derived inputs have larger amplitude during the movement period - especially for the second movement of the sequence. In addition, the data show trigger-like signals before each of the two reaches. Finally, while the data show a very high correlation between the pattern of preparatory activity of the second reach in the double reach and compound reach conditions, the derived inputs appear to be more different between the two conditions. Note that the data would be consistent with separate planning of the two reaches even in the compound reach condition, as well as the re-use of the preparatory input between the compound and double reach conditions. Therefore, different motor sequence datasets - notably, those that would show even more coarticulation between submovements - may be more promising to find a tight match between the data and the author's inputs. Further analyses in these datasets could help determine whether the coarticulation could be due to simple filtering by the circuits and muscles downstream of M1, planning of movements with adjusted curvature to mitigate the work performed by the muscles while permitting some amount of re-use across different sequences, or - as suggested by the authors - inputs fully tailored to one specific movement sequence that maximize accuracy and minimize the M1 input magnitude.

      Regarding the exact shape of the occupancy plots, it is important to note that some of the more qualitative aspects (e.g the relative height of the two peaks) will change if we change the parameters of the cost function. Right now, we have chosen the parameters to ensure that both reaches would be performed at roughly the same speed (as a way to very loosely constrain the parameters based on the observed behavior). However, small changes to the hyperparameters can lead to changes in the model output (e.g one of the two consecutive reaches being performed using greater acceleration than the other), and since our biophysical model is fairly simple, changes in the behavior are directly reflected in the network activity. Essentially, what this means is that while the double occupancy is a consistent feature of the model, the exact shape of the peaks is more sensitive to hyperparameters, and we do not wish to draw any strong conclusions from them, given the simplicity of the biophysical model. However, we do agree that our model exhibits some differences with the data. As discussed above, we have included additional discussion regarding the potential existence of separate inputs for planning vs triggering the movement in the context of single reaches.

      Overall, we are excited about the suggestions made by the Reviewer here about using our approach to analyze other motor sequence datasets, but we think that in order to do this properly, one would need to adopt a more realistic musculo-skeletal model (such as one provided by MotorNet).

      (4) Though iLQR is a powerful optimization method to find inputs optimizing the author's cost function, it also has some limitations. First, given that it relies on a linearization of the dynamics at each timestep, it has a limited ability to leverage potential advantages of nonlinearities in the dynamics. Second, the iLQR algorithm is not a biologically plausible learning rule and therefore it might be difficult for the brain to learn to produce the inputs that it finds. It remains unclear whether using alternative algorithms with different limitations - for instance, using variants of BPTT to train a separate RNN to produce the inputs in question - could impact some of the results.

      We agree that our choice of iLQR has limitations: while it offers the advantage of convergence guarantees, it does indeed restrict the choice of cost function and dynamics that we can use. We have now included extensive discussion of how the modeling choices affect our results.

      We do not view the lack of biological plausibility of iLQR as an issue, as the results are agnostic to the algorithm used for optimization. However, we agree that any structure imposed on the inputs (e.g by enforcing them to be the output of a self-contained dynamical system) would likely alter the results. A potentially interesting extension of our model would be to do just what the reviewer suggested, and try to learn a network that can generate the optimal inputs. However, this is outside the scope of our investigation, as it would then lead to new questions (e.g what brain region would that other RNN represent?).

      (5)  Under the objective considered by the authors, the amount of input occurring before the movement might be impacted by the presence of online sensory signals for closed-loop control. It is therefore an open question whether the objective and network characteristics suggested by the authors could also explain the presence of preparatory activity before e.g. grasping movements that are thought to be more sensory-driven (Meirhaeghe et al., Cell Reports 2023).

      It is true that we aren’t currently modeling sensory signals explicitly. However, some of the optimal inputs we infer may be capturing upstream information which could englobe some sensory information. This is currently unclear, and would likely depend on how exactly the model is specified. We have added new discussion to emphasize that our dynamics should not be understood as just representing M1, but more general circuits whose state can be decoded from M1.

      Reviewer #2 (Recommendations For The Authors):

      Additionally, thank you for pointing out various typos in the manuscript, we have fixed those!

      Reviewer 3:

      Thank you very much for your review, which makes a lot of very insightful points, and raises several interesting questions. In summary, we very much agree with the limitations you pointed out. In particular, the choice of input cost is something we had previously discussed, but we had found it challenging to decide on what a reasonable cost for “complexity” could be. Following your comment, we have however added a first attempt at penalizing “temporal complexity”, which shows promising behavior. We have only included those additional analyses as supplementary figures, and we have included new discussion, which hopefully highlights what we meant by the different model components, and how the model behavior may change as we vary some of our choices. We hope this can be informative for future models that may use a similar approach. Below, we highlight the changes that we have made to address your comments.

      The main limitation of the study is that it focuses exclusively on one specific constraint - magnitude - that could limit motor-cortex inputs. This isn't unreasonable, but other constraints are at least as likely, if less mathematically tractable. The basic results of this study will probably be robust with regard such issues - generally speaking, any constraint on what can be delivered during execution will favor the strategy of preparing - but this robustness cuts both ways. It isn't clear that the constraint used in the present study - minimizing upstream energy costs - is the one that really matters. Upstream areas are likely to be limited in a variety of ways, including the complexity of inputs they can deliver. Indeed, one generally assumes that there are things that motor cortex can do that upstream areas can't do, which is where the real limitations should come from. Yet in the interest of a tractable cost function, the authors have built a system where motor cortex actually doesn't do anything that couldn't be done equally well by its inputs. The system might actually be better off if motor cortex were removed. About the only thing that motor cortex appears to contribute is some amplification, which is 'good' from the standpoint of the cost function (inputs can be smaller) but hardly satisfying from a scientific standpoint.

      The use of a term that punishes the squared magnitude of control signals has a long history, both because it creates mathematical tractability and because it (somewhat) maps onto the idea that one should minimize the energy expended by muscles and the possibility of damaging them with large inputs. One could make a case that those things apply to neural activity as well, and while that isn't unreasonable, it is far from clear whether this is actually true (and if it were, why punish the square if you are concerned about ATP expenditure?). Even if neural activity magnitude an important cost, any costs should pertain not just to inputs but to motor cortex activity itself. I don't think the authors really wish to propose that squared input magnitude is the key thing to be regularized. Instead, this is simply an easily imposed constraint that is tractable and acts as a stand-in for other forms of regularization / other types of constraints. Put differently, if one could write down the 'true' cost function, it might contain a term related to squared magnitude, but other regularizing terms would by very likely to dominate. Using only squared magnitude is a reasonable way to get started, but there are also ways in which it appears to be limiting the results (see below).

      I would suggest that the study explore this topic a bit. Is it possible to use other forms of regularization? One appealing option is to constrain the complexity of inputs; a long-standing idea is that the role of motor cortex is to take relatively simple inputs and convert them to complex time-evolving inputs suitable for driving outputs. I realize that exploring this idea is not necessarily trivial. The right cost-function term is not clear (should it relate to low-dimensionality across conditions, or to smoothness across time?) and even if it were, it might not produce a convex cost function. Yet while exploring this possibility might be difficult, I think it is important for two reasons.

      First, this study is an elegant exploration of how preparation emerges due to constraints on inputs, but at present that exploration focuses exclusively on one constraint. Second, at present there are a variety of aspects of the model responses that appear somewhat unrealistic. I suspect most of these flow from the fact that while the magnitude of inputs is constrained, their complexity is not (they can control every motor cortex neuron at both low and high frequencies). Because inputs are not complexity-constrained, preparatory activity appears overly complex and never 'settles' into the plateaus that one often sees in data. To be fair, even in data these plateaus are often imperfect, but they are still a very noticeable feature in the response of many neurons. Furthermore, the top PCs usually contain a nice plateau. Yet we never get to see this in the present study. In part this is because the authors never simulate the situation of an unpredictable delay (more on this below) but it also seems to be because preparatory inputs are themselves strongly time-varying. More realistic forms of regularization would likely remedy this.

      That is a very good point, and it mirrors several concerns that we had in the past. While we did focus on the input norm for the sake of simplicity, and because it represents a very natural way to regularize our control solutions, we agree that a “complexity cost” may be better suited to models of brain circuits. We have addressed this in a supplementary investigation. We chose to focus on a cost that penalizes the temporal complexity of the inputs, as ||u(t+1) - u(t)||^2. Note that this required augmenting the state of the model, making the computations quite a bit slower; while it is doable if we only penalize the first temporal derivative, it would not scale well to higher orders.

      Interestingly, we did find that the activity in that setting was somewhat more realistic (see new Supplementary Figure S8), with more sustained inputs and plateauing activity. While we have kept the original model for most of the investigations, the somewhat more realistic nature of the results under that setting suggests that further exploration of penalties of that sort could represent a promising avenue to improve the model.

      We also found the idea of a cost that would ensure low-dimensionality of the inputs across conditions very interesting. However, it is challenging to investigate with iLQR as we perform the optimization separately for each condition; nevertheless, it could be investigated using a different optimizer.

      At present, it is also not clear whether preparation always occurs even with no delay. Given only magnitude-based regularization, it wouldn't necessarily have to be. The authors should perform a subspace-based analysis like that in Figure 6, but for different delay durations. I think it is critical to explore whether the model, like monkeys, uses preparation even for zero-delay trials. At present it might or might not. If not, it may be because of the lack of more realistic constraints on inputs. One might then either need to include more realistic constraints to induce zero-delay preparation, or propose that the brain basically never uses a zero delay (it always delays the internal go cue after the preparatory inputs) and that this is a mechanism separate from that being modeled.

      I agree with the authors that the present version of the model, where optimization knows the exact time of movement onset, produces a reasonably realistic timecourse of preparation when compared to data from self-paced movements. At the same time, most readers will want to see that the model can produce realistic looking preparatory activity when presented with an unpredictable delay. I realize this may be an optimization nightmare, but there are probably ways to trick the model into optimizing to move soon, but then forcing it to wait (which is actually what monkeys are probably doing). Doing so would allow the model to produce preparation under the circumstances where most studies have examined it. In some ways this is just window-dressing (showing people something in a format they are used to and can digest) but it is actually more than that, because it would show that the model can produce a reasonable plateau of sustained preparation. At present it isn't clear it can do this, for the reasons noted above. If it can't, regularizing complexity might help (and even if this can't be shown, it could be discussed).

      In summary, I found this to be a very strong study overall, with a conceptually timely message that was well-explained and nicely documented by thorough simulations. I think it is critical to perform the test, noted above, of examining preparatory subspace activity across a range of delay durations (including zero) to see whether preparation endures as it does empirically. I think the issue of a more realistic cost function is also important, both in terms of the conceptual message and in terms of inducing the model to produce more realistic activity. Conceptually it matters because I don't think the central message should be 'preparation reduces upstream ATP usage by allowing motor cortex to be an amplifier'. I think the central message the authors wish to convey is that constraints on inputs make preparation a good strategy. Many of those constraints likely relate to the fact that upstream areas can't do things that motor cortex can do (else you wouldn't need a motor cortex) and it would be good if regularization reflected that assumption. Furthermore, additional forms of regularization would likely improve the realism of model responses, in ways that matter both aesthetically and conceptually. Yet while I think this is an important issue, it is also a deep and tricky one, and I think the authors need considerable leeway in how they address it. Many of the cost-function terms one might want to use may be intractable. The authors may have to do what makes sense given technical limitations. If some things can't be done technically, they may need to be addressed in words or via some other sort of non-optimization-based simulation.

      Specific comments

      As noted above, it would be good to show that preparatory subspace activity occurs similarly across delay durations. It actually might not, at present. For a zero ms delay, the simple magnitude-based regularization may be insufficient to induce preparation. If so, then the authors would either have to argue that a zero delay is actually never used internally (which is a reasonable argument) or show that other forms of regularization can induce zero-delay preparation.

      Yes, that is a very interesting analysis to perform, which we had not considered before! When investigating this, we found that the zero-delay strategy does not rely on preparation in the same way as is seen in the monkeys. This seems to be a reflection of the fact that our “Go cue” corresponds to an “internal” go cue which would likely come after the true, “external go cue” – such that we would indeed never actually be in the zero delay setting. This is not something we had addressed (or really considered) before, although we had tried to ensure we referred to “delta prep” as the duration of the preparatory period but not necessarily the delay period. We have now included more discussion on this topic, as well as a new Supplementary Figure S10.

      I agree with the authors that prior modeling work was limited by assuming the inputs to M1, which meant that prior work couldn't address the deep issue (tackled here) of why there should be any preparatory inputs at all. At the same time, the ability to hand-select inputs did provide some advantages. A strong assumption of prior work is that the inputs are 'simple', such that motor cortex must perform meaningful computations to convert them to outputs. This matters because if inputs can be anything, then they can just be the final outputs themselves, and motor cortex would have no job to do. Thus, prior work tried to assume the simplest inputs possible to motor cortex that could still explain the data. Most likely this went too far in the 'simple' direction, yet aspects of the simplicity were important for endowing responses with realistic properties. One such property is a large condition-invariant response just before movement onset. This is a very robust aspect of the data, and is explained by the assumption of a simple trigger signal that conveys information about when to move but is otherwise invariant to condition. Note that this is an implicit form of regularization, and one very different from that used in the present study: the input is allowed to be large, but constrained to be simple. Preparatory inputs are similarly constrained to be simple in the sense that they carry only information about which condition should be executed, but otherwise have little temporal structure. Arguably this produces slightly too simple preparatory-period responses, but the present study appears to go too far in the opposite direction. I would suggest that the authors do what they can to address these issue via simulations and/or discussion. I think it is fine if the conclusion is that there exist many constraints that tend to favor preparation, and that regularizing magnitude is just one easy way of demonstrating that. Ideally, other constraints would be explored. But even if they can't be, there should be some discussion of what is missing - preparatory plateaus, a realistic condition-invariant signal tied to movement onset - under the present modeling assumptions.

      As described above, we have now included two additional figures. In the first one (S8, already discussed above), we used a temporal smoothness prior, and we indeed get slightly more realistic activity plateaus. In a second supplementary figure (S9), we have also considered using model predictive control (MPC) to optimize the inputs under an uncertain go cue arrival time. There, we found that removing the assumption that the delay period is known came with new challenges: in particular, it requires the specification of a “mental model” of when the Go cue will arrive. While it is reasonable to expect that monkeys will have a prior over the go time arrival cue that will be shaped by the design of the experiment, some assumptions must be made about the utility functions that should be used to weigh this prior. For instance, if we imagine that monkeys carry a model of the possible arrival time of the go cue that is updated online, they could nonetheless act differently based on this information, for instance by either preparing so as to be ready for the earliest go cue possible or alternatively to be ready for the average go cue. This will likely depend on the exact task design and reward/penalty structure. Here, we added simulations with those two cases (making simplifying assumptions to make the problem tractable/solvable using model predictive control), and found that the “earliest preparation” strategy gives rise to more realistic plateauing activity, while the model where planning is done for the “most likely go time” does not. We suspect that more realistic activity patterns could be obtained by e.g combining this framework with the temporal smoothness cost. However, the main point we wished to make with this new supplementary figure is that it is possible to model the task in a slightly more realistic way (although here it comes at the cost of additional model assumptions). We have now added more discussion related to those points. Note that we have kept our analyses on these new models to a minimum, as the main takeaway we wish to convey from them is that most components of the model could be modified/made more realistic. This would impact the qualitative behavior of the system and match to data but – in the examples we have so far considered – does not appear to modify the general strategy of networks relying on preparation.

      On line 161, and in a few other places, the authors cite prior work as arguing for "autonomous internal dynamics in M1". I think it is worth being careful here because most of that work specifically stated that the dynamics are likely not internal to M1, and presumably involve inter-area loops and (at some latency) sensory feedback. The real claim of such work is that one can observe most of the key state variables in M1, such that there are periods of time where the dynamics are reasonably approximated as autonomous from a mathematical standpoint. This means that you can estimate the state from M1, and then there is some function that predicts the future state. This formal definition of autonomous shouldn't be conflated with an anatomical definition.

      Yes, that is a good point, thank you for making it so clearly! Indeed, as previous work, we do not think of our “M1 dynamics” as being internal to M1, but they may instead include sensory feedback / inter-area loops, which we summarize into the connectivity, that we chose to have dynamics that qualitatively resemble data. We have now incorporated more discussion regarding what exactly the dynamics in our model represent.

      Round 2 of reviews

      Reviewer 3:

      My remaining comments largely pertain to some subtle (but to me important) nuances at a few locations in the text. These should be easy for the authors to address, in whatever way they see fit.

      Specific comments:

      (1) The authors state the following on line 56: "For preparatory processes to avoid triggering premature movement, any pre-movement activity in the motor and dorsal pre-motor (PMd) cortices must carefully exclude those pyramidal tract neurons."

      This constraint is overly restrictive. PT neurons absolutely can change their activity during preparation in principle (and appear to do so in practice). The key constraint is looser: those changes should have no net effect on the muscles. E.g., if d is the vector of changes in PT neuron firing rates, and b is the vector of weights, then the constraint is that b'd = 0. d = 0 is one good way of doing this, but only one. Half the d's could go up and half could go down. Or they all go up, but half the b's are negative. Put differently, there is no reason the null space has to be upstream of the PT neurons. It could be partly, or entirely, downstream. In the end, this doesn't change the point the authors are making. It is still the case that d has to be structured to avoid causing muscle activity, which raises exactly the point the authors care about: why risk this unless preparation brings benefits? However, this point can be made with a more accurate motivation. This matters, because people often think that a null-space is a tricky thing to engineer, when really it is quite natural. With enough neurons, preparing in the null space is quite simple.

      That is a good point – we have now reformulated this sentence to instead say “to avoid triggering premature movement, any pre-movement activity in the motor and dorsal premotor (PMd) cortices must engage the pyramidal tract neurons in a way that ensures their activity patterns will not lead to any movement”.

      (2) Line 167: 'near-autonomous internal dynamics in M1'.

      It would be good if such statements, early in the paper, could be modified to reflect the fact that the dynamics observed in M1 may depend on recurrence that is NOT purely internal to M1. A better phrase might be 'near-autonomous dynamics that can be observed in M1'. A similar point applies on line 13. This issue is handled very thoughtfully in the Discussion, starting on line 713. Obviously it is not sensible to also add multiple sentences making the same point early on. However, it is still worth phrasing things carefully, otherwise the reader may have the wrong impression up until the Discussion (i.e. they may think that both the authors, and prior studies, believe that all the relevant dynamics are internal to M1). If possible, it might also be worth adding one sentence, somewhere early, to keep readers from falling into this hole (and then being stuck there till the Discussion digs them out).

      That is a good point: we have now edited the text after line 170 to make it clear that the underlying dynamics may not be confined to M1, and have referenced the later discussion there.

      (3) The authors make the point, starting on line 815, that transient (but strong) preparatory activity empirically occurs without a delay. They note that their model will do this but only if 'no delay' means 'no external delay'. For their model to prepare, there still needs to be an internal delay between when the first inputs arrive and when movement generating inputs arrive.

      This is not only a reasonable assumption, but is something that does indeed occur empirically. This can be seen in Figure 8c of Lara et al. Similarly, Kaufman et al. 2016 noted that "the sudden change in the CIS [the movement triggering event] occurred well after (~150 ms) the visual go cue... (~60 ms latency)" Behavioral experiments have also argued that internal movement-triggering events tend to be quite sluggish relative to the earliest they could be, causing RTs to be longer than they should be (Haith et al. Independence of Movement Preparation and Movement Initiation). Given this empirical support, the authors might wish to add a sentence indicating that the data tend to justify their assumption that the internal delay (separating the earliest response to sensory events from the events that actually cause movement to begin) never shrinks to zero.

      While on this topic, the Haith and Krakauer paper mentioned above good to cite because it does ponder the question of whether preparation is really necessary. By showing that they could get RTs to shrink considerably before behavior became inaccurate, they showed that people normally (when not pressured) use more preparation time than they really need. Given Lara et al, we know that preparation does always occur, but Haith and Krakauer were quite right that it can be very brief. This helped -- along with neural results -- change our view of preparation from something more cognitive that had to occur, so something more mechanical that was simply a good network strategy, which is indeed the authors current point. Working a discussion of this into the current paper may or may not make sense, but if there is a place where it is easy to cite, it would be appropriate.

      This is a nice suggestion, and we thank the reviewer for pointing us to the Haith and Krakauer paper. We have now added this reference and extended the paragraph following line 815 to briefly discuss the possible decoupling between preparation and movement initiation that is shown in the Haith paper, emphasizing how this may affect the interpretation of the internal delay and comparisons with behavioral experiments.

    1. were probably three or four three or four suspicious actors and 20 suspicious actors total yes in that room

      Wait a second-- what exactly makes them "suspicious"? What she means by suspicious of course--without saying it directly-- is the idea that perhaps there were plants or some kind of dark agents that were provoking the crowd to do something other than what they would have done "naturally". But no evidence is offered for this.

      She may be referring to the widely covered mention in the Jan 6 report of "unindicted co-conspirators" but Andrew McCabe former acting director and deputy director of the FBI under Trump lays it out.

      “When an indictment is written, at the time the indictment is written, and signed off by the judge, there are people who need to be referred to, in the indictment, just to make it a coherent story, so it makes sense, but who the government is not prepared to charge at that time,” McCabe said. “There may be all kinds of different reasons that they’re not prepared to charge that person at that time.”

      “But the one reason that does not exist is the one that [Carlson] suggested,” McCabe added. “It’s not an undercover officer, because you cannot refer to those people as unindicted co-conspirators.”

      Indeed, federal case law from 1985 (United States v. Rodriguez) acknowledged that “government agents and informers cannot be conspirators.”

      Also, a Department of Justice manual advises federal prosecutors not to identify unindicted co-conspirators by name without “some significant justification.”

      In a 2004 paper, American University Washington College of Law Professor Ira P. Robbins argued that naming those individuals who have not been charged with a crime should be prohibited because it violates their due process rights.

    2. let alone have the identity the name of the police officer involved have his name concealed from the public for months um that just never happens

      This simply isn't true. Its totally routine for an officer's name to be withheld pending an internal investigation. Further, how is the "media" to blame as being in "cahoots" with congress even related to this. It's actually the DC police that were in charge of the investigation.

    3. this agent or suspicious actor

      Again, sure he's got tactical gear on and a face mask-- but of course maybe he's just a guy that likes to contribute to chaos and doesn't want to get caught. He could be a Proud Boy that's trying to stir the pot. He could be anyone. There's no evidence either way.

      Ultimately the protestors are responsible for their own actions. If I see that the capitol is being attacked my first response isn't going to be to climb into a broken window, it's going to be to get the hell out of there and stay away so that the authorities can get things under control.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews: 

      Reviewer #1 (Public Review): 

      Summary: 

      In this work, Odenwald and colleagues show that mutant biotin ligases used to perform proximity-dependent biotin identification (TurboID) can be used to amplify signal in fluorescence microscopy and to label phase-separated compartments that are refractory to many immunofluorescence approaches. Using the parasite Trypanosoma brucei, they show that fluorescent methods such as expansion microscopy and CLEM, which require bright signals for optimal detection, benefit from the elevated signal provided by TurboID fusion proteins when coupled with labeled streptavidin. Moreover, they show that phase-separated compartments, where many antibody epitopes are occluded due to limited diffusion and potential sequestration, are labeled reliably with biotin deposited by a TurboID fusion protein that localizes within the compartment. They show successful labeling of the nucleolus, likely phase-separated portions of the nuclear pore, and stress granules. Lastly, they use a panel of nuclear pore-TurboID fusion proteins to map the regions of the T. brucei nuclear pore that appear to be phase-separated by comparing antibody labeling of the protein, which is susceptible to blocking, to the degree of biotin deposition detected by streptavidin, which is not. 

      Strengths: 

      Overall, this study shows that TurboID labelling and fluorescent streptavidin can be used to boost signal compared to conventional immunofluorescence in a manner similar to tyramide amplification, but without having to use antibodies. TurboID could prove to be a viable general strategy for labeling phase-separated structures in cells, and perhaps as a means of identifying these structures, which could also be useful. 

      Weaknesses: 

      However, I think that this work would benefit from additional controls to address if the improved detection that is being observed is due to the increased affinity and smaller size of streptavidin/biotin compared to IgGs, or if it has to do with the increased amount of binding epitope (biotin) being deposited compared to the number of available antibody epitopes. I also think that using the biotinylation signal produced by the TurboID fusion to track the location of the fusion protein and/or binding partners in cells comes with significant caveats that are not well addressed here, mostly due to the inability to discern which proteins are contributing to the observed biotin signal. 

      To dissect the contributions of the TurboID fusion to elevating signal, anti-biotin antibodies could be used to determine if the abundance of the biotin being deposited by the TurboID is what is increasing detection, or if streptavidin is essential for this.

      We agree with the reviewer, that it would be very interesting to distinguish whether the increase in signal comes from the multiple biotinylation sites or from streptavidin being a very good binder, or perhaps from both. However, this question is very hard to answer, as antibodies differ massively in their affinity to the antigen which is further dependent on the respective IF-conditions, and are therefore not directly comparible. Even if anti-biotin gives a better signal then anti-HA, this can be either caused by the increase in antigen-number (more biotin than HA-tag) or by the higher binding affinity, or by a combination of both, thus hard to distinguish. Nevertheless, we have tested monoclonal mouse anti-biotin targeting the (non-phase-separated) NUP158. We found the signal from the biotin-antibody to be much weaker than from anti-HA, indicating that, at least this particular biotin antibody, is not a very good binder in IF. 

      Alternatively, HaloTag or CLIP tagging could be used to see if diffusion of a small molecule tag other than biotin can overcome the labeling issue in phase-separated compartments. There are Halo-biotin substrates available that would allow the conjugation of 1 biotin per fusion protein, which would allow the authors to dissect the relative contributions of the high affinity of streptavidin from the increased amount of biotin that the TurboID introduces. 

      This is a very good idea, as in this case, the signals are both from streptavidin and are directly comparable. We expressed NUP158 with HaloTag and added PEG-biotin as a Halo ligand. However, PEG-biotin is poorly cell-permeable, and is in general only used on lysates. In trypanosomes, cell permeability is particular restricted, and even Halo-ligands that are considered highly cell-penetrant give only a weak signal. Even after over-night incubation, we could not get any signal with PEG-biotin. Our control, the TMR-ligand 647, gave a weak nuclear pore staining, confirming the correct expression and function of the HaloTag-NUP158.

      The idea of using the biotin signal from the TurboID fusion as a means to track the changing localization of the fusion protein or the location of interacting partners is an attractive idea, but the lack of certainty about what proteins are carrying the biotin signal makes it very difficult to make clear statements. For example, in the case of TurboID-PABP2, the appearance of a biotin signal at the cell posterior is proposed to be ALPH1, part of the mRNA decapping complex. However, because we are tracking biotin localization and biotin is being deposited on a variety of proteins, it is not formally possible to say that the posterior signal is ALPH1 or any other part of the decapping complex. For example, the posterior labeling could represent a localization of PABP2 that is not seen without the additional signal intensity provided by the TurboID fusion. There are also many cytoskeletal components present at the cell posterior that could be being biotinylated, not just the decapping complex. Similar arguments can be made for the localization data pertaining to MLP2 and NUP65/75. I would argue that the TurboID labeling allows you to enhance signal on structures, such as the NUPs, and effectively label compartments, but you lack the capacity to know precisely which proteins are being labeled.  

      We fully agree with the reviewer, that tracking proteins by streptavidin imaging alone is problematic, because it cannot distinguish, which protein is biotinylated. We therefore used words like “likely”  in the description of the data. However, we still think, it is a valid method, as long as it is confirmed by an orthogonal method. We have added this paragraph to the end of this chapter:

      “Importantly, tracking of proteins by streptavidin imaging requires orthogonal controls, as the imaging alone does not provide information about the nature of the biotinylated proteins. These can be proximity ligation assay, mass spectrometry or specific tagging visualisation of protein suspects by fluorescent tags. Once these orthogonal controls are established for a specific tracking, streptavidin imaging is an easy and cheap and highly versatile method to monitor protein interactions in a specific setting.”

      Reviewer #2 (Public Review): 

      Summary: 

      The authors noticed that there was an enhanced ability to detect nuclear pore proteins in trypanosomes using a streptavidin-biotin-based detection approach in comparison to conventional antibody-based detection, and this seemed particularly acute for phase-separated proteins. They explored this in detail for both standard imaging but also expansion microscopy and CLEM, testing resolution, signal strength, and sensitivity. An additional innovative approach exploits the proximity element of biotin labelling to identify where interacting proteins have been as well as where they are. 

      Strengths: 

      The data is high quality and convincing and will have obvious application, not just in the trypanosome field but also more broadly where proteins are tricky to detect or inaccessible due to phase separation (or some other steric limitations). It will be of wide utility and value in many cell biological studies and is timely due to the focus of interest on phase separation, CLEM, and expansion microscopy. 

      Thank you! We are glad you liked it.

      Reviewer #3 (Public Review): 

      Summary: 

      The authors aimed to investigate the effectiveness of streptavidin imaging as an alternative to traditional antibody labeling for visualizing proteins within cellular contexts. They sought to address challenges associated with antibody accessibility and inconsistent localization by comparing the performance of streptavidin imaging with a TurboID-HA tandem tag across various protein localization scenarios, including phase-separated regions. They aimed to assess the reliability, signal enhancement, and potential advantages of streptavidin imaging over antibody labeling techniques. 

      Overall, the study provides a convincing argument for the utility of streptavidin imaging in cellular protein visualization. By demonstrating the effectiveness of streptavidin imaging as an alternative to antibody labeling, the study offers a promising solution to issues of accessibility and localization variability. Furthermore, while streptavidin imaging shows significant advantages in signal enhancement and preservation of protein interactions, the authors must consider potential limitations and variations in its application. Factors such as the fact that tagging may sometimes impact protein function, background noise, non-specific binding, and the potential for off-target effects may impact the reliability and interpretation of results. Thus, careful validation and optimization of streptavidin imaging protocols are crucial to ensure reproducibility and accuracy across different experimental setups. 

      Strengths: 

      - Streptavidin imaging utilizes multiple biotinylation sites on both the target protein and adjacent proteins, resulting in a substantial signal boost. This enhancement is particularly beneficial for several applications with diluted antigens, such as expansion microscopy or correlative light and electron microscopy. 

      - This biotinylation process enables the identification and characterization of interacting proteins, allowing for a comprehensive understanding of protein-protein interactions within cellular contexts. 

      Weaknesses: 

      - One of the key advantages of antibodies is that they label native, endogenous proteins, i.e. without introducing any genetic modifications or exogenously expressed proteins. This is a major difference from the approach in this manuscript, and it is surprising that this limitation is not really mentioned, let alone expanded upon, anywhere in the manuscript. Tagging proteins often impacts their function (if not their localization), and this is also not discussed.

      - Given that BioID proximity labeling encompasses not only the protein of interest but also its entire interacting partner history, ensuring accurate localization of the protein of interest poses a challenge. 

      - The title of the publication suggests that this imaging technique is widely applicable. However, the authors did not show the ability to track the localization of several distinct proteins on the same sample, which could be an additional factor demonstrating the outperformance of streptavidin imaging compared with antibody labeling. Similarly, the work focuses only on small 2D samples. It would have been interesting to be able to compare this with 3D samples (e.g. cells encapsulated in an extracellular matrix) or to tissues.  

      Recommendations for the authors:

      To enhance the assessment from 'incomplete' to 'solid', the reviewers recommend that the following major issues be addressed: 

      Major issues: 

      (1) Anti-biotin antibodies in combination with TurboID labeling should be used to compare the signal/labelling penetrance to streptavidin results. That would show if elevated biotin deposition matters, or if it is really the smaller size, more fluors, and higher affinity of streptavidin that's making the difference. 

      We agree with the reviewer, that it would be very interesting to distinguish whether the increase in signal comes from the multiple biotinylation sites or from streptavidin being a very good binder, or perhaps from both, and whether the size matters (IgG versus streptavidin). However, this question is very hard to answer, as antibodies differ massively in their affinity to the antigen. Thus, even if antibiotin would give a better signal then anti-HA, this could be either caused by the increase in antigen-number (more biotin than HA-tag) or by the better binding affinity, or by a combination, and it would not allow to truly answer the question. We have now tested anti-biotin antibodies, also in repsonse to reviewer 1, and got a much poorer signal in comparison to anti-HA or streptavidin.

      Please note that we made another attempt using nanobodies to target phase-separated proteins, to see, whether size matters (Fig. 2I). The nanobody did not stain Mex67 at the nuclear pores, but gave a weak nucelolar signal for NOG1, which may suggest that the nanobody can slightly better penetrate than IgG, but it does not rule out that the nanobody simply binds with higher affinity. Reviewer 1 has suggested to use the Halo Tag with PEG-biotin: this would indeed allow to directly compare the streptavidin signal caused by the TurboID with a single biotin added by the Halo tag. Unfortunately, the PEG-biotin does not  penetrate trypanosome cells. In conclusion, we are not aware of a method that would allow to establish why streptavidin but not IgGs can penetrate to phase separated areas. We therefore prefer to not overinterpret our data, but stick to what is supported by the data: “the inability to label phase-separated areas is not restricted to anti-HA but applies to other antibodies”.

      (3) Figure 4 A-B. The validity of claiming the correct localization demonstrated by streptavidin imaging comes into question, especially when endogenous fluorescence, via the fusion protein, remains undetectable (as indicated by the yellow arrow at apex). 

      In this figure, the streptavidin imaging does NOT show the correct localisation of the bait protein, but it does show proteins from historic interactions that have a distinct localisation to the bait. We had therefore introduced this chapter with the paragraph below, to make sure, the reader is aware of the limitations (which we also see as an opportunity, if properly controlled):

      “We found that in most cases, streptavidin labelling faithfully reflects the steady state localisation of a bait protein, e.g., the localisation resembles those observed with immunofluorescence or direct fluorescence imaging of GFP-fusion proteins. For certain bait proteins, this is not the case, for example, if the bait protein or its interactors have a dynamic localisation to distinct compartments, or if interactions are highly transient. It is thus essential to control streptavidin-based de novo localisation data by either antibody labelling (if possible) or by direct fluorescence of fusion-proteins for each new bait protein.”

      In particular, on lines 450-460, there's a fundamental issue with the argument put forward here. It is not possible to formally know that the posterior labeling is ALPH1 vs. another part of the decapping complex that was associated with PABP2-Turbo, or if the higher detection capacity of the Turbo-biotin label is uncovering a novel localization of the PABP2. While it is likely that it is ALPH1, it is not possible to rule out other possibilities with this approach. These issues should be discussed here and more generally the possibility of off-target labeling with this approach should be addressed in the discussion. 

      We fully agree with the reviewer, that tracking proteins by streptavidin imaging alone is problematic, because it cannot distinguish, which protein is biotinylated. We therefore used words like “likely”  in the description of the data. However, we still think, it is a valid method, as long as it is back-uped by an orthogonal method. We have added this paragraph to the end of this chapter:

      “Importantly, tracking of proteins by streptavidin imaging requires orthogonal controls, as the imaging alone does not provide information about the nature of the biotinylated proteins. These can be proximity ligation assay, mass spectrometry or specific tagging visualisation of protein suspects by fluorescent tags. Once these orthogonal controls are established for a specific tracking, streptavidin imaging is an easy and cheap and highly versatile method to monitor protein interactions in a specific setting.”

      (4) More discussion and acknowledgment of the general limitations in using tagged proteins are needed to balance the manuscript, especially if the hope is to draw a comparison with antibody labeling, which works on endogenous proteins (not requiring a tag). For example: (a) tagging proteins requires genetic/molecular work ahead of time to engineer the constructs and/or cells if trying to tag endogenous proteins; (b) tagged proteins should technically be validated in rescue experiments to confirm the tag doesn't disrupt function in the cell/tissue/context of interest; and (c) exogenous tagged proteins compete with endogenous untagged proteins, which can complicate the interpretation of data.  

      We have added this paragraph to the first paragraph of the discussion part:

      “Like many methods that are frequently used in cell- and molecular biology, streptavidin imaging is based on the expression of a genetically engineered fusion protein: it is essential to validate both, function and localisation of the TurboID-HA tagged protein by orthogonal methods. If the fusion protein is non-functional or mis-localised, tagging at the other end may help, but if not, this protein cannot be imaged by streptavidin imaging. Likewise, target organisms not amenable to genetic manipulation, or those with restricted genetic tools,  are not or less suitable for this method.”

      Also, we like to point out that for non-mainstream organisms like trypanosomes, antibodies are not commercially available and often genetic manipulation is more time-efficient and cheaper than the production of antiserum against the target protein.

      Also, the introduction would ideally be more general in scope and introduce the pros and cons of antibody labeling vs biotin/streptavidin, which are mentioned briefly in the discussion. The fact that the biotin-streptavidin interaction is ~100-fold higher affinity than an IgG binding to its epitope is likely playing a key role in the results here. The difference in size between IgG and streptavidin, the likelihood that the tetrameric streptavidin carries more fluors than a IgG secondary, and the fact that biotin can likely diffuse into phase-separated environments should be clearly stated. The current introduction segues from a previous paper that a more general audience may not be familiar with. 

      We have now included this paragraph to the introduction:

      “It remains unclear, why streptavidin was able to stain biotinylated proteins within these antibody inaccessible regions, but possible reasons are: (i) tetrameric streptavidin is smaller and more compact than IgGs (60 kDa versus a tandem of two IgGs, each with 150 kDa) (ii) the interaction between streptavidin and biotin is ~100 fold stronger than a typical interaction between antibody and antigen and (iii) streptavidin contains four fluorophores, in contrast to only one per secondary IgG.”

      Minor issues: 

      The copy numbers of the HA and Ty1 epitope tags vary depending on the construct being used. For example, Ty1 is found as a single copy tag in the TurboID tag, but on the mNeonGreen tag there are 6 copies of the epitope. It makes it hard to know if differences in detection are due to variations in copies of the epitope tags. Line 372-374: can the authors explain why they chose to use nanobodies in this case? It would be great to show the innate mNeonGreen signal in 2K to compare to the Ty1 labeling. The presence of 6 copies of the Ty1 epitope could be essential to the labeling seen here.

      We agree with the reviewer, that these data are a bit confusing. We have now removed Figure 3K, as it is the only construct with 6 Ty1 instead of one, and it does not add to the conclusions. (the mNeonsignal is entirely in the nucleolus, as shown by Tryptag). We have also added an explanation why we used nanobodies (“The absence of a nanobody signal rules out that its simply the size of IgGs that prevents the staining of Mex67 at the nuclear pores, as nanobodies are smaller than (tetrameric) streptavidin”). However, as stated above, we prefer not to overinterpret the data, as signals from different antibodies/nanobodies – antigen combinations are not comparable. Important to us was to stress that the absence of signal in phase-separated areas is NOT restricted to the anti-HA antibody, which is clearly supported by the data.

      What is the innate streptavidin background labeling look like in cells that are not carrying a TurboID fusion, from the native proteins that are biotinylated? That should be discussed. 

      We have now included the controls without the TurboID fusions for trypanosomes and HeLa cells: “Wild type cells of both Trypanosomes and human showed only a very low streptavidin signal, indicating that the signal from naturally biotinylated proteins is neglectable (Figure S8 in supplementary material).”

      Line 328-331: This is likely to be dependent on whether or not the protein moves to different localizations within the cell. 

      True, we agree, and we have added this paragraph:

      “The one exception are very motile proteins that produce a “biotinylation trail” distinct to the steady state localisation; these exceptions, and how they can be exploited to understand protein interactions, are discussed in chapter 4 below. “

      Line 304-305: Does biotin supplementation not matter at all? 

      No, we never saw any increase in biotinylation when we added extra biotin to trypanosomes. The 0.8 µM biotin concentration in the medium were sufficient.

      Line 326-327: Was the addition of biotin checked for enhancement in the case of the mammalian NUP98? I would argue that there is a significant number of puncta in Figure 1D that are either green or magenta, not both. The amount of extranuclear puncta in the HA channel is also difficult to explain. Biotin supplementation to 500 µM was used in mammalian TurboID experiments in the original Nature Biotech paper- perhaps nanomolar levels are too low. 

      We now tested HeLa cells with 500 µM Biotin and saw an increase in signal, but also in background; due to the increased background  we conclude that low biotin concentrations are more suitable . We have also repeated the experiment using 4HA tags instead of 1HA, and we found a minor improvement in the antibody signal for NUP88 (while the phase separated NUP54 was still not detectable). We have replaced the images in Figure 1D  (NUP88) and also in Figure 2F (NUP54) with improved images and using 4HA tags. However, we like to note that single nuclear pore resolution is beyond what can be expected of light microscopy.

      Line 371: In 2I, I see a signal that looks like the nucleus, similar to the Ty1 labeling in 2G, so I don't think it's accurate to say that that Mex67 was "undetectable". Does the serum work for blotting? 

      Thank you, yes, “undetectable” was not the correct phrase here. Mex67 localises to the nuclear pores, to the nuceoplasm and to the nucleolus (GFP-tagging or streptavidin). Antibodies, either to the tag or to the endogenous proteins, fail to detect Mex67 at the nuclear pores and also don’t show any particular enrichment in the nucleolus. They do, however, detect Mex67 in the (not-phase-separated) area of the nucleoplasm. We have changed the text to make this clearer. The Mex67 antiserum works well on a western blot (see for example: Pozzi, B., Naguleswaran, A., Florini, F., Rezaei, Z. & Roditi, I. The RNA export factor TbMex67 connects transcription and RNA export in Trypanosoma brucei and sets boundaries for RNA polymerase I. Nucleic Acids Res. 51, 5177–5192 (2023))

      Line 477: "lacked" should be "lagged".

      Thank you, corrected.

      Line 468-481: My previous argument holds here - how do you know that the difference in detection here is just a matter of much higher affinity/quantity of binding partner for the avidin?

      See answer to the second point of (3), above.

      483-491: Same issue - without certainty about what the biotin is on, this argument is difficult to make. 

      See answer to the second point of (3), above.

      Line 530: "bone-fine" should be "bonafide"

      Thank you, corrected.

      Line 602: biotin/streptavidin labeling has been used for expansion microscopy previously (Sun, Nature Biotech 2021; PMID: 33288959). 

      Thank you, we had overlooked this! We have now included this reference and describe the differences to our approach clearer in the discussion part:

      “Fluorescent streptavidin has been previously used in expansion microscopy to detect biotin residues in target proteins produced by click chemistry (Sun et al., 2021). However, to the best of our knowledge, this is the first report that employs fluorescent streptavidin as a signal enhancer in expansion microscopy and CLEM, by combining it with multiple biotinylation sites added by a biotin ligase. Importantly, for both CLEM and expansion, streptavidin imaging is the only alternative approach to immunofluorescence, as denaturing conditions associated with these methods rule out direct imaging of fluorescent tags.”

    1. Reviewer #2 (Public Review):

      Summary:

      This study uses an elegant design, using cross-decoding of multivariate fMRI patterns across different types of stimuli, to convincingly show a functional dissociation between two sub-regions of the parietal cortex, the anterior inferior parietal lobe (aIPL) and superior parietal lobe (SPL) in visually processing actions. Specifically, aIPL is found to be sensitive to the causal effects of observed actions (e.g. whether an action causes an object to compress or to break into two parts), and SPL to the motion patterns of the body in executing those actions.

      To show this, the authors assess how well linear classifiers trained to distinguish fMRI patterns of response to actions in one stimulus type can generalize to another stimulus type. They choose stimulus types that abstract away specific dimensions of interest. To reveal sensitivity to the causal effects of actions, regardless of low-level details or motion patterns, they use abstract animations that depict a particular kind of object manipulation: e.g. breaking, hitting, or squashing an object. To reveal sensitivity to motion patterns, independently of causal effects on objects, they use point-light displays (PLDs) of figures performing the same actions. Finally, full videos of actors performing actions are used as the stimuli providing the most complete, and naturalistic information. Pantomime videos, with actors mimicking the execution of an action without visible objects, are used as an intermediate condition providing more cues than PLDs but less than real action videos (e.g. the hands are visible, unlike in PLDs, but the object is absent and has to be inferred). By training classifiers on animations, and testing their generalization to full-action videos, the classifiers' sensitivity to the causal effect of actions, independently of visual appearance, can be assessed. By training them on PLDs and testing them on videos, their sensitivity to motion patterns, independent of the causal effect of actions, can be assessed, as PLDs contain no information about an action's effect on objects.

      These analyses reveal that aIPL can generalize between animations and videos, indicating that it is sensitive to action effects. Conversely, SPL is found to generalize between PLDs and videos, showing that it is more sensitive to motion patterns. A searchlight analysis confirms this pattern of results, particularly showing that action-animation decoding is specific to right aIPL, and revealing an additional cluster in LOTC, which is included in subsequent analyses. Action-PLD decoding is more widespread across the whole action observation network.

      This study provides a valuable contribution to the understanding of functional specialization in the action observation network. It uses an original and robust experimental design to provide convincing evidence that understanding the causal effects of actions is a meaningful component of visual action processing and that it is specifically localized in aIPL and LOTC.

      Strengths:

      The authors cleverly managed to isolate specific aspects of real-world actions (causal effects, motion patterns) in an elegant experimental design, and by testing generalization across different stimulus types rather than within-category decoding performance, they show results that are convincing and readily interpretable. Moreover, they clearly took great care to eliminate potential confounds in their experimental design (for example, by carefully ordering scanning sessions by increasing realism, such that the participants could not associate animation with the corresponding real-world action), and to increase stimulus diversity for different stimulus types. They also carefully examine their own analysis pipeline, and transparently expose it to the reader (for example, by showing asymmetries across decoding directions in Figure S3). Overall, this is an extremely careful and robust paper.

      Weaknesses:

      I list several ways in which the paper could be improved below. More than 'weaknesses', these are either ambiguities in the exact claims made, or points that could be strengthened by additional analyses. I don't believe any of the claims or analyses presented in the paper show any strong weaknesses, problematic confounds, or anything that requires revising the claims substantially.

      (1) Functional specialization claims: throughout the paper, it is not clear what the exact claims of functional specialization are. While, as can be seen in Figure 3A, the difference between action-animation cross-decoding is significantly higher in aIPL, decoding performance is also above chance in right SPL, although this is not a strong effect. More importantly, action-PLD cross-decoding is robustly above chance in both right and left aIPL, implying that this region is sensitive to motion patterns as well as causal effects. I am not questioning that the difference between the two ROIs exists - that is very convincingly shown. But sentences such as "distinct neural systems for the processing of observed body movements in SPL and the effect they induce in aIPL" (lines 111-112, Introduction) and "aIPL encodes abstract representations of action effect structures independently of motion and object identity" (lines 127-128, Introduction) do not seem fully justified when action-PLD cross-decoding is overall stronger than action-animation cross-decoding in aIPL. Is the claim, then, that in addition to being sensitive to motion patterns, aIPL contains a neural code for abstracted causal effects, e.g. involving a separate neural subpopulation or a different coding scheme? Moreover, if sensitivity to motion patterns is not specific to SPL, but can be found in a broad network of areas (including aIPL itself), can it really be claimed that this area plays a specific role, similar to the specific role of aIPL in encoding causal effects? There is indeed, as can be seen in Figure 3A, a difference between action-PLD decoding in SPL and aIPL, but based on the searchlight map shown in Figure 3B I would guess that a similar difference would be found by comparing aIPL to several other regions. The authors should clarify these ambiguities.

      (2) Causal effect information in PLDs: the reasoning behind the use of PLD stimuli is to have a condition that isolates motion patterns from the causal effects of actions. However, it is not clear whether PLDs really contain as little information about action effects as claimed. Cross-decoding between animations and PLDs is significant in both aIPL and LOTC, as shown in Figure 4. This indicates that PLDs do contain some information about action effects. This could also be tested behaviorally by asking participants to assign PLDs to the correct action category. In general, disentangling the roles of motion patterns and implied causal effects in driving action-PLD cross-decoding (which is the main dependent variable in the paper) would strengthen the paper's message. For example, it is possible that the strong action-PLD cross-decoding observed in aIPL relies on a substantially different encoding from, say, SPL, an encoding that perhaps reflects causal effects more than motion patterns. One way to exploratively assess this would be to integrate the clustering analysis shown in Figure S1 with a more complete picture, including animation-PLD and action-PLD decoding in aIPL.

      (3) Nature of the motion representations: it is not clear what the nature of the putatively motion-driven representation driving action-PLD cross-decoding is. While, as you note in the Introduction, other regions such as the superior temporal sulcus have been extensively studied, with the understanding that they are part of a feedforward network of areas analyzing increasingly complex motion patterns (e.g. Riese & Poggio, Nature Reviews Neuroscience 2003), it doesn't seem like the way in which SPL represents these stimuli are similarly well-understood. While the action-PLD cross-decoding shown here is a convincing additional piece of evidence for a motion-based representation in SPL, an interesting additional analysis would be to compare, for example, RDMs of different actions in this region with explicit computational models. These could be, for example, classic motion energy models inspired by the response characteristics of regions such as V5/MT, which have been shown to predict cortical responses and psychophysical performance both for natural videos (e.g. Nishimoto et al., Current Biology 2011) and PLDs (Casile & Giese Journal of Vision 2005). A similar cross-decoding analysis between videos and PLDs as that conducted on the fMRI patterns could be done on these models' features, obtaining RDMs that could directly be compared with those from SPL. This would be a very informative analysis that could enrich our knowledge of a relatively unexplored region in action recognition. Please note, however, that action recognition is not my field of expertise, so it is possible that there are practical difficulties in conducting such an analysis that I am not aware of. In this case, I kindly ask the authors to explain what these difficulties could be.

      (4) Clustering analysis: I found the clustering analysis shown in Figure S1 very clever and informative. However, there are two things that I think the authors should clarify. First, it's not clear whether the three categories of object change were inferred post-hoc from the data or determined beforehand. It is completely fine if these were just inferred post-hoc, I just believe this ambiguity should be clarified explicitly. Second, while action-anim decoding in aIPL and LOTC looks like it is consistently clustered, the clustering of action-PLD decoding in SPL and LOTC looks less reliable. The authors interpret this clustering as corresponding to the manual vs. bimanual distinction, but for example "drink" (a unimanual action) is grouped with "break" and "squash" (bimanual actions) in left SPL and grouped entirely separately from the unimanual and bimanual clusters in left LOTC. Statistically testing the robustness of these clusters would help clarify whether it is the case that action-PLD in SPL and LOTC has no semantically interpretable organizing principle, as might be the case for a representation based entirely on motion pattern, or rather that it is a different organizing principle from action-anim, such as the manual vs. bimanual distinction proposed by the authors. I don't have much experience with statistical testing of clustering analyses, but I think a permutation-based approach, wherein a measure of cluster robustness, such as the Silhouette score, is computed for the clusters found in the data and compared to a null distribution of such measures obtained by permuting the data labels, should be feasible. In a quick literature search, I have found several papers describing similar approaches: e.g. Hennig (2007), "Cluster-wise assessment of cluster stability"; Tibshirani et al. (2001) "Estimating the Number of Clusters in a Data Set Via the Gap Statistic". These are just pointers to potentially useful approaches, the authors are much better qualified to pick the most appropriate and convenient method. However, I do think such a statistical test would strengthen the clustering analysis shown here. With this statistical test, and the more exhaustive exposition of results I suggested in point 2 above (e.g. including animation-PLD and action-PLD decoding in aIPL), I believe the clustering analysis could even be moved to the main text and occupy a more prominent position in the paper.

      (5) ROI selection: this is a minor point, related to the method used for assigning voxels to a specific ROI. In the description in the Methods (page 16, lines 514-24), the authors mention using the MNI coordinates of the center locations of Brodmann areas. Does this mean that then they extracted a sphere around this location, or did they use a mask based on the entire Brodmann area? The latter approach is what I'm most familiar with, so if the authors chose to use a sphere instead, could they clarify why? Or, if they did use the entire Brodmann area as a mask, and not just its center coordinates, this should be made clearer in the text.

    1. reply to u/IndividualCoast9039 at https://new.reddit.com/r/typewriters/comments/1endi5d/screenwriter_here/

      There's really no such thing as a screenplay specific machine, though for ease of use, you'll surely want one with a tabulator (tabs). If you want to hew toward the standard screenplay formatting look for pica machines (10 characters per inch) rather than elite machines (11-12 characters per inch).

      SoCal is lousy with lots of great machines. If you want something that's going to work "out of the box" you'll pay a few bones more, but unless you're a tinkerer, it's definitely worth it.

      I'd recommend checking out the following shops/repair joints near LA that specialize in machines for writers. Most will let you try out the touch and feel of a few in person to figure out what will work best for you. Putting your hands on actual machines will help you know which one you'll want for yourself.

      • Helmut Schulze, Rees Electronics / Star Typewriters, 2140 Westwood Blvd. #224, Los Angeles, CA 90025. 310-475-0859 or 877-219-1450. Fax: 310-475-0850. E-mail star@startypewriters.com. Schulze has many years of experience and has restored typewriters of famous writers for collector Steve Soboroff.
      • Aaron Therol @ Typewriter Connection, DTLA, https://www.typewriterconnection.com/
      • Bob Marshall, Typewriter Muse, Riverside, CA. Service, restoration, and sales. Website: typewritermuse.com.
      • Rubin Flores at U.S. Office Machine Co. over in Highland Park 323-256-2111 (better at repairs, restoration; I don't think he keeps stock)

      I'd generally endorse most of the advice on models you'll find in these sources which are geared specifically toward writers, all three sources have lots experience and reasonable bona fides to make such recommendations.

      All machines are slightly different, so pick the one that speaks to you and your methods of working.

      If it helps to know what typewriters actual (screen) writers have used in the past, check out https://site.xavier.edu/polt/typewriters/typers.html

      Beyond this Just My Typewriter has a few short videos that'll give you a crash course on Typewriter 101: https://www.youtube.com/playlist?list=PLJtHauPh529XYHI5QNj5w9PUdi89pOXsS

    1. The kind that let their eyes feather across the titles like trailing fingertips, heads cocked, with book-hunger rising off them like heatwaves from July pavement.

      This is a really interesting and descriptive way to describe people who like to read rather than just saying "they like to read". It gives a clearer image, makes it much more interesting to read, and it's something this author uses well and frequently throughout.

    1. Hegel claims a religious point of view can help furnish individuals with ethical principles and help us lead a more ethical life

      Yeah but is this really saying that it's the way it ought to be, or is it just saying that religion provides a framework to make determinations.

      I mean, one thing you could argue is that by hastening the process of making determinations, we ought to be religious because it'll help us reach the next determination.

    1. Stop selling courses the hard way, just because it's what everyone tells you to do. Finally do something about the parts of selling that cause stress or overload. Trust in the value of what you do to boldly share it with the world to make more money

      shorten to one: the opposite of stop hiding

    1. Welcome back.

      This is part two of this lesson. We're gonna continue immediately from the end of part one, so let's get started.

      Right at the offset, it's critical to understand that layer 2 of the networking stack uses layer 1. So in order to have an active layer 2 network, you need the layer 1 or physical layer in place and working, meaning two network interfaces and a shared medium between those as a minimum. But now on top of this, each networking stack, so remember this is all of the networking software layers, each stack, so the left and the right, now have both layer 1 and layer 2. So conceptually, layer 2 sits on top of layer 1 like this, and our games, these are now communicating using layer 2.

      Each device, because it's layer 2, now has a MAC address, a hardware address, which is owned by the network interface card on each device. So let's step through what happens now when the game on the left wants to send something to the game running on the right. Well, the left game knows the MAC address of the game on the right. Let's assume that as a starting point. So it communicates with the layer 2 software. Let's assume this is ethernet on the left. It indicates that it wants to send data to the MAC address on the right and the layer 2 software based on this, creates an ethernet frame containing the data that the game wants to send in the payload part of the frame.

      So the frame F1 has a destination of the MAC address ending 5b76, which is the MAC address on the laptop on the right and it contains within the payload part of that frame the data that the game wants to send. At this point, this is where the benefits of layer 2 begin. Layer 2 can communicate with the layer 1 part of the networking stack and it can look for any signs of a carrier signal. If any other device on the network were transmitting at this point, you would see the signal on the layer 1 network. So it's looking to sense a carrier, and this is the job of CSMA, which stands for carrier sense multiple access. In this case, it doesn't detect a carrier and so it passes the frame to layer 1. Layer 1 doesn't care what the frame is, it doesn't understand the frame as anything beyond a block of data and so it transmits this block of data onto the shared medium.

      On the right side, the layer 1 software receives the raw bit stream and it passes it up to its layer 2. Layer 2 reviews the destination MAC address of the frame. It sees that the frame is destined for itself and so it can pass that payload, that data, back to the game and so that's how these games can communicate using layer 2. So layer 2 is using layer one to transmit and receive the raw data, but it's adding on top of this MAC addresses which allow for machine-to-machine communication and in addition, it's adding this media access control.

      So let's look at an example of how this works. If we assume that at the same time that the left machine was transmitting, the machine on the right attempted to do the same. So layer 2 works with layer 1 and it checks for a carrier on the shared physical medium. If the left machine is transmitting, which we know it is, the carrier is detected and layer 2 on the right simply waits until the carrier is not detected. So it's layer 2 which is adding this control. Layer 1 on its own would simply transmit and cause a collision, but layer 2 checks for any carrier before it instructs layer 1 to transmit. When the carrier isn't detected anymore, then layer 2 sends the frame down to layer 1 for transmission. Layer 1 just sees it as data to send and it transmits it across the physical medium. It's received at the left side device. Layer 1 sends this raw data to layer 2. It can see that it's the intended destination and so it sends the data contained in the frame payload back to the game.

      Now, I want to reiterate a term again which you need to remember, and that term is encapsulation. This is the process of taking some data, in this case game data, and wrapping it in something else. In this case, the game data is encapsulated inside a frame at each side before giving the game back its intended data, the data is de-encapsulated, the payload is extracted from the frame. So this is a concept that you need to remember, because as data passes down the OSI model, it's encapsulated in more and more different components. So the transport layer does some encapsulation, the network layer does some encapsulation and the data link layer does some encapsulation. This is a process which you need to be comfortable with.

      Now, two more things which I want to cover on this screen and the first of this is conceptually, for anything using layer 2 to communicate, they see it as layer 2 on the left is directly communicating to layer 2 on the right. Even though layer 2 is using layer 1 to perform the physical communication, anything which is using these layer 2 services has no visibility of that. That's something that's common in the OSI model. Anything below the point that you are communicating with is abstracted away. If you're using a web browser which functions at layer 7, you don't have to worry about how your data gets to the web server. It just works. So your web browser, which is running at layer 7 is communicating with a web server which is also running at layer 7. You don't have to worry about how this communication happens.

      The other thing which I want to cover, is if in this scenario, what if both machines check for a carrier which doesn't exist and then both layer 2s instruct their layer 1 to both transmit at the same time. This causes a collision. Now, layer 2 contains collision detection and that's what the CD part of CSMA/CD is for. If a collision is detected, then a jam signal is sent by all of the devices which detect it and then a random back-off occurs. The back-off is a period of time during which no device will attempt a transmission. So after this back-off period occurs, the transmission is retried. Now, because this back-off is random, hopefully it means that only one device will attempt to transmit at first and other devices will see the carrier on the network and wait before transmitting, but if we still have a collision, then this back-off is attempted again only now with a greater period. So over time, there's less and less likelihood that you're going to have multiple devices transmitting at the same time. So this collision detection and avoidance is essential for layer 2. It's the thing that allows multiple devices to coexist on the same layer 2 network.

      Okay, so let's move on.

      So now you have an idea about layer 2 networking, let's revisit how our laptops are connected. In the previous example where I showed hubs, we had four devices connected to the same four port hub. Now, hubs are layer 1 devices. This means they don't understand frames in any way. They just see physical data. Essentially, they're a multi-port repeater. They just repeat any physical activity on one port to all or the ports. So the top laptop sends a frame destined for the bottom laptop. The hub just sees this as raw data. It repeats this on all of the other ports and this means all of the other laptops will receive this data, which their layer 2 software will interpret as a frame. They'll all receive this data, so this frame, they'll see that they're not the intended destination and they'll discard it. The bottom laptop will see that it is the intended destination and its layer 2 software will pass on the data to the game.

      Now, hubs aren't smart and this means that if the laptop on the right were to start transmitting at exactly the same time, then it would cause a collision, and this collision would be repeated on all of the other ports that the hub has connected. Using a layer 1 device, a hub doesn't prevent you running a layer 2 network over the top of it, but it means that the hub doesn't understand layer 2 and so it behaves in a layer 1 way. You still have each device doing carrier sense multiple access, and so collisions should be reduced, but if multiple devices try and transmit at exactly the same time, then this will still cause collisions.

      What will improve this is using a switch, and a switch is a layer 2 device. It works in the same way physically as a hub, but it understands layer 2 and so it provides significant advantages. Let's review how this changes things. To keep things simple, let's keep the same design only now we have a switch in the middle. It still has four ports and it's still connected to the same four laptops. Because they're layer 2, they now have their own hardware addresses, their MAC addresses. Now, because a switch is a layer 2 device, it means that it has layer 2 software running inside it, which means that it understands layer 2. And because of that it maintains what's called a MAC address table. Switches over time learn what's connected to each port, so the device MAC addresses which are connected to each port. When a switch sees frames, it can interpret them and see the source and destination MAC addresses. So over time with this network, the MAC address table will get populated with each of our devices. So the switch will store the MAC addresses it sees on a port and the port itself. This is generally going to happen the first time each of the laptop sends a frame which the switch receives. It will see the source MAC address on the frame and it will update the MAC address table which it maintains.

      Now let's say the MAC address table is populated and the top laptop sends a frame which is intended for the left laptop. Well, the switch will see the frame arrive at the port that the top laptop is connected to at which point, one of two things will happen. If the switch didn't know which port the destination MAC address was on, well it would forward this frame to all of the other ports. If it does know which port the specific MAC address is attached to, then it will use that one port to forward the frame to. Switches are intelligent. They aren't just repeating the physical level. They interpret the frames and make decisions based on the source and destination MAC addresses of those frames. So switches store and forward frames. They receive the frame, they store it and then they forward it based on the MAC address table and then they discard it. Now this has another benefit, because it's not just repeating like a dumb layer 1 device, it means that it won't forward collisions. In fact, each port on the switch is a separate collision domain. Because each port is a separate collision domain, the only two things which can transmit at the same time are the device and the port it's connected to. So if there is a collision, it will be limited to that one port only. The switch will not forward that corrupted data through to any of its other ports, because it only forwards valid frames. The switch isn't forwarding the physical layer, it's dealing with frames only. It receives a frame. If it's valid, it stores it, it reviews it and then it forwards it.

      Now, layer 2 is the foundation for all networks which you use day-to-day. It's how your wired networks work. It's how wifi networks work. It's how the internet works, which is basically a huge collection of interconnected layer 2 networks. The name itself, so the internet, stands for an inter-network of networks. So these networks are layer 2 networks which are all connected together to form the internet.

      So in summary, what position are we in by adding layer 2?

      Well, we have identifiable devices. For the first time, we can uniquely address frames to a particular device using MAC addresses. It allows for device to device communication rather than the shared media which layer 1 offers. We also have media access control so devices can share access to physical media in a nice way avoiding crosstalk and collisions. But we also have the ability to detect collisions and so if they do occur, we have a way to correct or work around them. So with all of that, we can do unicast communications which are one-to-one communications and we can do broadcast communications, which are one-to-all communications. And as long as we replace layer 1 hubs with layer two switches, which are like hubs with superpowers, then we gain the ability to scale to a much better degree and we avoid a lot of the collisions because switches store and forward rather than just repeating everything.

      Now, layer 2 really does provide crucial functionality. Everything from this point onwards builds on layer 2, so it's critical that you understand it. So if necessary, then re-watch this video, because from now on, you need to understand layer 2 at a real fundamental level. Now, this seems like a great point to take a break, so go ahead and complete this video and when you're ready, join me in the next part of this series where we'll be looking at layer 3, which is the network layer.

    1. Welcome back.

      And in this part of the lesson series, we're going to look at layer two of the OSI model, which is the data link layer.

      Now the data link layer is one of the most critical layers in the entire OSI seven-layer model. Everything above this point relies on the device-to-device communication, which the data link layer provides. So when you are sending or receiving data, to or from the internet, just be aware that the data link layer is supporting the transfer of that data. So it's essential that you understand.

      Now, this is going to be one of the longer parts of this lesson series, because layer two actually provides a significant amount of functionality.

      Now, before I step through the architecture of layer two, we have to start with the fundamentals. Layer two, which is the data link layer, runs over layer one. So a layer two network requires a functional layer one network to operate, and that's something which is common throughout the OSI model. Higher layers build on lower layers adding features and capabilities. A layer two network can run on different types of layer one networks, so copper, fiber, wifi, and provide the same capabilities.

      Now there are different layer two protocols and standards for different situations, but for now, we're going to focus on ethernet which is what most local networks use. So things in your office or things in your home.

      Now, layer two, rather than being focused on physical wavelengths or voltages, introduces the concept of frames. And frames are a format for sending information over a layer two network. Layer two also introduces a unique hardware address, called a MAC address, for every device on a network. This hardware address is a hexadecimal address. It's 48 bits long and it looks like this: 3e:22 and so on. The important thing to understand is that a MAC address, generally, for physical networking, is not software assigned. The address is uniquely attached to a specific piece of hardware.

      A MAC address is formed of two parts, the OUI, which is the organizationally unique identifier, and this is assigned to companies who manufacture network devices. So each of these companies will have a separate OUI. The second part of the MAC address is then network interface controller or NIC specific. And this means together the MAC address on a network card should be globally unique.

      Now, layer two, as I've mentioned, uses layer one. This means that a layer two, or ethernet frame, can be transmitted onto the shared physical medium by layer one. This means that it's converted into voltages, RF, or light. It's sent across the medium and then received by other devices, also connected to that shared medium.

      It's important to understand this distinction. Layer two provides frames, as well as other things which I'll cover soon. And layer one handles the physical transmission and reception onto and from the physical shared medium. So when layer one is transmitting a frame onto the physical medium, layer one doesn't understand the frame. Layer one is simply transmitting raw data onto that physical medium.

      Now a frame, which is the thing that layer two uses for communication, is a container of sorts. It has a few different components. The first part is the preamble and start frame delimiter. And the function of this is to allow devices to know that it's the start of the frame, so they can identify the various parts of that frame. You need to know where the start of a frame is to know where the various parts of that frame start.

      Next comes the destination and the source MAC addresses. So all devices on a layer two network have a unique MAC address, and a frame can be sent to a specific device on a network by putting its MAC address in the destination field, or you can put all Fs if you want to send the frame to every device on the local network. And this is known as a broadcast. Now the source MAC address field is set to the device address of whatever is transmitting the frame and this allows it to receive replies.

      Next is EtherType, and this is commonly used to specify which layer three protocol is putting its data inside a frame. Just like layer two uses layer one to move raw bitstream data across the shared physical medium, while layer three uses layer two frames for device-to-device communication on a local network. And so when you are receiving a frame at the other side of a communication, you need to know which layer three protocol originally put data into that frame. A common example might be IP or the internet protocol. So this is what the EtherType, or ET, field is used for.

      Now, these three fields are commonly known as the MAC header. They control the frame destination, they indicate the source and specify its function. After the header is the payload, and this is anywhere from 46 to 1500 bytes in size for standard frames. It contains the data that the frame is sending. The data is generally provided by the layer three protocol and the protocol which is being used, as I just mentioned, is indicated within the EtherType or ET field.

      Now this process is called encapsulation. You have something which layer three generates, often this is an IP Packet, and this is put inside an ethernet frame. It's encapsulated in that frame. The frame delivers that data to a different layer two destination. On the other side, the frame is analyzed, and the layer three packet is extracted and given back to layer three at the destination side. The EtherType field is used to determine which layer three protocol receives this at the destination. And then finally at the end of the frame, is the frame check sequence, which is used to identify any errors in the frame. It's a simple CRC check. It allows the destination to check if corruption has occurred or not.

      So that's the frame and it's an important thing to understand if you are to understand how all of the bits of layer one, two, and three fit together. So layer two frames are generated by the layer two software at the source side. They're passed to layer one. That raw data is transmitted onto the physical medium. It's taken off the physical medium at the destination side. It's passed to its layer two software and that can interpret the frame and pass that onto layer three, which can then interpret that data.

      Now as a reminder, this is the problem that we have with a purely layer one network implementation. We have two devices running a game, a laptop on the left and a laptop on the right. And these are connected using a single network cable, a shared physical medium. Now, as I mentioned earlier in this lesson series, layer one provides no media access control. The layer one software rolling on the network card will simply transmit any data it receives onto the physical medium. So if the game on the left sends some data, it will be transmitted onto the medium and it will be seen by the device on the right. The problem is that the laptop on the right could also be sending data at the same time. This means the electrical signals will overlap and interfere with each other and this is known as a collision and it impacts both pieces of data. It corrupts both.

      This is one of the problems of layer one, which is solved by layer two, and layer two provides controlled access to the physical medium. Now let's explore how.

      Okay, so this is the end of part one of this lesson. It was getting a little bit on the long side and so I wanted to add a break. It's an opportunity just to take a rest or grab a coffee. Part two will be continuing immediately from the end of part one. So go ahead, complete the video, and when you're ready, join me in part two.

    1. Welcome to this lesson where I'm going to be talking about high availability (HA), fault tolerance (FT), and disaster recovery (DR). It's essential that you understand all three of these to be an effective solutions architect and I want to make sure that you understand all of them correctly. Many of the best architects and consultants that I've worked with have misunderstood exactly what HA and FT mean. The best outcome of this misunderstanding is that you waste business funds and put a project at risk. Worst case, you can literally put lives at risk. So, let's jump in and get started and I promise to keep it as brief as possible, but this really is something you need to fully understand.

      Let's start with high availability. This is a term that most people think that they understand. Formally, the definition is that high availability aims to ensure an agreed level of operational performance, usually uptime, for a higher than normal period and I've highlighted the key parts of that definition. Most students that I initially teach have an assumption that making a system highly available means ensuring that the system never fails or that the user of a system never experiences any outages and that is not true. HA isn't aiming to stop failure, and it definitely doesn't mean that customers won't experience outages. A highly available system is one designed to be online and providing services as often as possible. It's a system designed so that when it fails, its components can be replaced or fixed as quickly as possible, often using automation to bring systems back into service. High availability is not about the user experience. If a system fails and a component is replaced and that disrupts service for a few seconds, that's okay. It's still highly available. High availability is about maximizing a system's online time and that's it.

      Let me give you an example. Let's say we have a system which has a customer, Winnie. Winnie is a data scientist and uses a bespoke application to identify complex data trends. Now, this application runs on a single server, let's say inside AWS. The application probably has other users in addition to Winnie. It's an important application to the business. If it's down, the staff can't work. If they can't work, they don't generate value to the business and of course, this costs the business money. If we have a failure, it means that the system is now suffering an outage, it's not available. System availability is generally expressed in the form of a percentage of uptime. So we might have 99.9 or three nines and this means that we can only have 8.77 hours of downtime per year. Imagine only being able to take a system down for 8.77 hours a year, that's less than one hour per month. It gets worse though, some systems need even higher levels of availability. We've got 99.999% availability or five nines and this only allows for 5.26 minutes per year of downtime. That means for all outages during a year, you have 5.26 minutes. That includes identifying that there's an outage, identifying the cause, devising a solution, and implementing a fix. An outage in this context is defined as something which impacts that server, so impacts your users.

      Now, fixing Winnie's application quickly can be done by swapping out the compute resource, probably a virtual server. Rather than using time to diagnose the issue, if you have a process ready to replace it, it can be fixed quickly and probably in an automated way, or you might improve this further by having two servers online constantly, one active and one standby. In the event of a failure, customers would move to the standby server with very close to zero downtime. But, and this is a key factor about high availability, when they migrate from the active server to the standby server, they might have to re-login or might have some small disruption. For high availability, user disruption, while not being ideal, is okay. It can happen because high availability is just about minimizing any outages.

      Now, this might explain it a little better. This is a real-world example of something which has high availability built in. It's a four by four. If you were driving in the desert with a normal urban-grade car and it got a flat tire, would you have a spare? Would you have the tools ready to repair it as quickly as possible? In a desert, an outage or delay could have major impacts. It's risky and it could impact getting to your destination. So an example of high availability is to carry a spare wheel and the tools required to replace it. You would of course, need to spend time changing the tire, which is a disruption, but it could be done and it minimizes the time that you're out of action. If you don't have a spare tire, then you'd need to call for assistance, which would substantially increase the time you're out of action. So, high availability is about keeping a system operational. It's about fast or automatic recovery of issues. It's not about preventing user disruption. While that's a bonus, a highly available system can still have disruption to your user base when there is a failure.

      Now, high availability has costs required to implement it. It needs some design decisions to be made in advance and it requires a certain level of automation. Sometimes, high availability needs redundant servers or redundant infrastructure to be in place ready to switch customers over to in the event of a disaster to minimize downtime.

      Now, let's take this a step further and talk about fault tolerance and how it differs from high availability. When most people think of high availability, they're actually mixing it up with fault tolerance. Fault tolerance in some ways is very similar to high availability, but it is much more. Fault tolerance is defined as the property that enables a system to continue operating properly in the event of a failure of some of its components, so one or more faults within the system. Fault tolerance means that if a system has faults, and this could be one fault or multiple faults, then it should continue to operate properly, even while those faults are present and being fixed. It means it has to continue operating through a failure without impacting customers.

      Imagine a scenario where we have somebody injured, so we've got Dr. Abbie and she's been told that she has an urgent case of an injured patient and we'll call this patient, Mike. Mike has been rushed to the hospital after injuring himself running. He's currently being prepped for a surgical procedure and is in the operating room and currently under general anesthetic. While he's unconscious, he's being monitored and this monitoring system indicates when to reduce or increase the levels of anesthetic that Mike gets. It's critical that this server is not to be interrupted ever. The system uses underlying infrastructure on-premises at the hospital. Now, in the event of a system failure, if it was just a highly available system, the server could be replaced or another server could be included in an active standby architecture. In either case, the swap between the servers would cause a system error, a disruption. However quick the fix, however small that disruption, in certain situations like this, any disruption can be life-threatening. This is an example of a situation where high availability isn't enough. Fault tolerance systems are designed to work through failure with no disruption. In this example, we might have the system's monitor communicating with two servers at the same time in an active, active configuration. The monitor is connected to both servers all of the time. So this is not just a simple fail-over configuration. If a server failed, it would drop down to just communicating with the remaining server and as long as one server remains active, the system is fully functional. Now, we could take this further adding a second monitoring system, itself with connections to both servers. That way, one monitor can fail, one server can fail and still the service would continue uninterrupted. We could even eliminate the human dependency in the system and add an extra surgeon, Dr. Abbie's twin.

      Most people think that HA means operating through failure, it's not. HA is just about maximizing uptime. Fault tolerance is what means to operate through failure. Fault tolerance can be expensive because it's much more complex to implement versus high availability. High availability can be accomplished by having spare equipment, so standby, physical or virtual components. As long as you automate things and have these spare components ready to go, you can minimize outages. With fault tolerance, it's about more than that. You first need to minimize outages, which is the same as HA, but then you also need to design the system to be able to tolerate the failure, which means levels of redundancy and system components, which can route traffic and sessions around any failed components.

      Now remember the example I used for high availability, the four by four in the desert. There are situations where we can't pull over to the side of the road and change a component. An example of this is a plane, which is in the air. A plane needs to operate through systems failure, so through an engine failure, for example. If an engine fails, the plane can't simply stop and effect repairs. So, a plane comes with more engines than it needs. It comes with duplicate electronic systems and duplicate hydraulic systems, so that when it has a problem, it just carries on running until it can safely land and effect repairs. AWS is no exception to this. Systems can be designed to only maximize uptime, which is high availability, or they can be designed for mission or life critical situations and so, designed to operate through that failure, which is fault tolerance.

      As a solutions architect, you need to understand what your customer requires. A customer might say that they need HA or fault tolerance while not understanding the difference. Fault tolerance is harder to design, harder to implement and costs much more. Implementing fault tolerance when you really needed high availability simply means you're wasting money. It costs more, and it takes longer to implement. But the reverse, implementing high availability when you need fault tolerance, means that you're potentially putting life at risk. A highly available plane is less than ideal. Understand the difference, if you don't, it can be disastrous.

      So, let's move on to the final concept, which is disaster recovery. The definition of disaster recovery is a set of policies, tools, and procedures to enable the recovery or continuation of vital technology infrastructure and systems following a natural or human-induced disaster. So, while high availability and fault tolerance are about designing systems to cope or operate through disaster, disaster recovery is about what to plan for and do when disaster occurs, which knocks out a system. So, if high availability and fault tolerance don't work, what then? What if your building catches fire, is flooded or explodes? Disaster recovery is a multiple-stage set of processes. So given a disaster, it's about what happens before, so the pre-planning and what happens afterwards, the DR process itself.

      The worst time for any business is recovering in the event of a major disaster. In that type of environment, bad decisions are made, decisions based on shock, lack of sleep, and fear of how to recover. So, a good set of DR processes need to preplan for everything in advance. Build a set of processes and documentation, plan for staffing and physical issues when a disaster happens. If you have a business premises with some staff, then part of a good DR plan might be to have a standby premises ready and this standby premises can be used in the event of a disaster. That way, done in advance, your staff unaffected by the disaster, know exactly where to go. You might need space for IT systems or you might use a cloud platform, such as AWS as a backup location, but in any case, you need the idea of a backup premises or a backup location that's ready to go in the event of a disaster.

      If you have local infrastructure, then make sure you have resilience. Make sure you have plans in place and ready during a disaster. This might be extra hardware sitting at the backup site ready to go, or it might be virtual machines or instances operating in a cloud environment ready when you need them. A good DR plan means taking regular backups, so this is essential. But the worst thing you can do is to store these backups at the same site as your systems, it's dangerous. If your main site is damaged, your primary data and your backups are damaged at the same time and that's a huge problem. You need to have plans in place for offsite backup storage. So, in the event of a disaster, the backups can be restored at the standby location. So, have the backups of your primary data offsite and ready to go and make sure that all of the staff know the location and the access requirements for these backups.

      Effective DR planning isn't just about the tech though, it's about knowledge. Make sure that you have copies of all your processes available. All your logins to key systems need to be available for the staff to use when they're at this standby site. Do this in advance and it won't be a chaotic process when an issue inevitably occurs. Ideally, you want to run periodic DR testing to make sure that you have everything you need and then if you identify anything missing, you can refine your processes and run the test again. If high availability is a four-by-four, if fault tolerance are the resilient systems on large planes, then effective DR processes are pilot or passenger ejection systems. DR is designed to keep the crucial and non-replaceable parts of your system safe, so that when a disaster occurs, you don't lose anything irreplaceable and can rebuild after the disaster. Historically, disaster recovery was very manual. Because of cloud and automation, DR can now be largely automated, reducing the time for recovery and the potential for any errors.

      As you go through the course, I'm going to help you understand how to implement high availability and fault tolerance systems in AWS using AWS products and services. So, you need to understand both of these terms really well and disaster recovery. So in summary, high availability is about minimizing any outages, so maximizing system availability. Fault tolerance extends this, building systems which operate through faults and failures. Don't confuse the two. Fault tolerance is much more complex and expensive. It takes a lot more time and effort to implement and manage. I'll help you as we go through the course by identifying how to implement systems which are highly available and how to implement systems which are fault tolerant. AWS provides products and services which help with both of those or just help with one or the other and you need to know the difference. Disaster recovery is how we recover. It's what we do when high availability and fault tolerance don't work and AWS also has many systems and features which help with disaster recovery and one of the things that the exam tests will be your knowledge of how quickly you can recover and how best to recover, given the various different products and services and I'll highlight all of this as we go through the course. At this point, that's everything I wanted to cover, so thanks for listening. Go ahead, complete this video and when you're ready, I'll see you in the next.

    1. Welcome back. In this lesson, I'm going to be covering something that will make complete sense by the end of the course. I'm introducing it now because I want you to be thinking about it whenever we're talking about AWS products and services. The topic is the shared responsibility model. The easiest way to explain this is visually, so let's jump in.

      Remember earlier in the course when I talked about the various different cloud service models? In each of these models, there were parts of the infrastructure stack that you were responsible for as the customer, and parts of the infrastructure stack that the vendor or provider were responsible for. With IaaS, for example, the company providing the IaaS product, so AWS in the case of EC2, they're responsible for the facilities, the AWS data centers, the infrastructure, so storage and networking, the servers, so EC2 hosts, and the hypervisor that allows physical hardware to be carved up into independent virtual machines. You as the customer manage the operating system, any containers, any run times, the data on the instance, the application, and any ways in which it interfaces with its customers. This is an example of a set of shared responsibilities. Part of the responsibilities lie with the vendor, and part lie with you as the customer.

      The AWS shared responsibility model is like that, only applying to the wider cloud platform from a security perspective. It's AWS' way of making sure that it's clear and that you understand fully which elements you manage and which elements it manages. At a high level, AWS are responsible for the security of the cloud. You as a customer are responsible for the security in the cloud. Now let's explore this in a little bit more detail because it will help you throughout the course and definitely for the exam.

      Now I've covered the AWS infrastructure at a high level in a previous lesson. AWS provides these to you as a service that you consume. So AWS are responsible for managing the security of the AWS regions, the Availability Zones, and the edge locations. So the hardware and security of the global infrastructure. You have no control over any of that and you don't need to worry about it. It's the "of the cloud" part, and so it's AWS' responsibility. The same holds true for the compute storage databases and networking which AWS also provide to you. AWS manage the security of those components. In addition, any software which assists in those services, AWS manage all of this part of the stack. So the hardware, the regions, the global network, the compute storage database, and networking services, and then any software that is used to provide that service, AWS manage that end-to-end.

      If you consume a service from AWS, they handle the provisioning and the security of that thing. So take EC2 as an example. The region and the Availability Zone that the instance run in, that's AWS' responsibility. The compute, the storage, the underlying databases and networking for that service, from a security perspective, that's AWS' responsibility. The software, so the user interface, the hypervisor, that's handled by AWS. Now you accept responsibility for the operating system upwards. What does that include? It means things like the client-side data encryption, integrity and authentication; server-side encryption; network traffic protection. If your application encrypts its data, you manage that. If your server uses SSL certificates, you manage those. If you encrypt server-to-server communications, then you also handle that. You're also responsible for the operating system, networking, and any local firewall configuration. You're responsible for applications, identity and access management to things that you will need to implement, manage and control. And then any customer data. So any data that runs in this stack, you need to manage it, secure it, and ensure that it's backed up.

      This might seem like a pretty abstract concept. You might be wondering, does it actually benefit you in the exam? I'd agree with you to a point. When I was doing my AWS studies, I actually didn't spend much time on the shared responsibility model. But what I found is when I sat the exam, I did feel as though it could have benefited me to start learning about it early on when I was first starting my studies. If you keep the shared responsibility in mind as we're going through the various different AWS products, you'll start building up an idea of which elements of that product AWS manage, and which elements you're responsible for. When it comes to deploying an EC2 instance into a VPC or using the Relational Database Service to deploy and manage a database inside a VPC, you need to know which elements of that you manage and which elements AWS manage.

      I'll be referring back to this shared responsibility model fairly often as we go through the course, so you build up this overview of which elements you need to worry about and which are managed by AWS. If possible, I would suggest that you either print out the shared responsibility model and put it on your desk as you're studying, or just make sure you've got a copy that you can refer back to. It becomes important to understand it at this high level. I'm not going to use any more of your time on this topic. I just wanted to introduce it. I promise you that I'm not going to be wasting your time by talking about things which don't matter. This will come in handy. This is definitely something that will help you answer some questions.

      That's all I wanted to cover for now. It's just a foundation, and I don't want to bore you with too much isolated theory. Try to keep this in mind as you go through the rest of the course. For now, this is the level of detail that you need. That's everything I wanted to cover. Go ahead, complete this lesson. When you're ready, move on to the next.

    1. we know from previous large transitions in history that you never change the world  by having everyone on board. You change the world by having large enough minorities  that can tip quite inert majority to move in the right direction

      for - social tipping points - quote - Johan Rockstrom

      quote - social tipping points - Johan Rockstrom - (see below) - We know from previous large transitions in history that - You never change the world by having everyone on board.. - You change the world by having large enough minorities<br /> - that can tip quite inert majority to move in the right direction. - When you look at the world of sustainability. - in many societies in the world, we are actually a double digit penetration - on sustainable solutions, - on people's awareness, - on willingness to even politically vote for green or, sustainable options. - So we're very close to that positive tipping point as well. - and that's another reason why it's not the moment to back down. - Now is the moment to just increase momentum.

    2. there are other tipping  points, like for example, lakes. that can flip over from, you  know, oxygen rich, fish rich, clear water lakes into these murky,  algal bloom dominated, anoxic states, dead states, based on nutrient loading  and overfishing, and that is a Oh, not from climate or temperature. Not anything, no, has  nothing to do with climate or temperature, it's just a, mismanagement,

      for - other types of tipping points - not climate but human mismanagement of resources

    1. Author response:

      eLife assessment

      This useful study examines the neural activity in the motor cortex as a monkey reaches to intercept moving targets, focusing on how tuned single neurons contribute to an interesting overall population geometry. The presented results and analyses are solid, though the investigation of this novel task could be strengthened by clarifying the assumptions behind the single neuron analyses, and further analyses of the neural population activity and its relation to different features of behaviour.

      Thanks for recognizing the content of our research, and please stay tuned for our follow-up studies on neural dynamics during interception.

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      This study addresses the question of how task-relevant sensory information affects activity in the motor cortex. The authors use various approaches to address this question, looking at single units and population activity. They find that there are three subtypes of modulation by sensory information at the single unit level. Population analyses reveal that sensory information affects the neural activity orthogonally to motor output. The authors then compare both single unit and population activity to computational models to investigate how encoding of sensory information at the single unit level is coordinated in a network. They find that an RNN that displays similar orbital dynamics and sensory modulation to the motor cortex also contains nodes that are modulated similarly to the three subtypes identified by the single unit analysis.

      Strengths:

      The strengths of this study lie in the population analyses and the approach of comparing single-unit encoding to population dynamics. In particular, the analysis in Figure 3 is very elegant and informative about the effect of sensory information on motor cortical activity. The task is also well designed to suit the questions being asked and well controlled.

      We appreciate these kind comments.

      It is commendable that the authors compare single units to population modulation. The addition of the RNN model and perturbations strengthen the conclusion that the subtypes of individual units all contribute to the population dynamics. However, the subtypes (PD shift, gain, and addition) are not sufficiently justified. The authors also do not address that single units exhibit mixed modulation, but RNN units are not treated as such.

      We’re sorry for not providing sufficient grounds to introduce the subtypes. We determined the PD shift, gain, and addition as pertinent subtypes based on classical cosine tuning model (Georgopoulos et al., 1982) and referred to some gain modulation studies (e.g. Pesaran et al. 2010, Bremner and Andersen, 2012). Here, we applied this subtype analysis as a criteria to identify the modulation in neuronal population rather than to sort neuron into distinct cell types. We will update Methods in the revised version of manuscript.

      Weaknesses:

      The main weaknesses of the study lie in the categorization of the single units into PD shift, gain, and addition types. The single units exhibit clear mixed selectivity, as the authors highlight. Therefore, the subsequent analyses looking only at the individual classes in the RNN are a little limited. Another weakness of the paper is that the choice of windows for analyses is not properly justified and the dependence of the results on the time windows chosen for single-unit analyses is not assessed. This is particularly pertinent because tuning curves are known to rotate during movements (Sergio et al. 2005 Journal of Neurophysiology).

      The mixed selectivity or precisely the mixed modulation is indeed a significant feature of neuronal population in the present study. The purpose of the subtype analysis was to serve as a criterion for the potential modulation mechanisms. However, the results appear to be a spectrum than clusters. It still through some insights to understand the modulation distribution and we will refine the description in the next version. In the current version, we observed single-unit tuning and population neural state with sliding windows, focusing on the period around movement onset (MO) due to the emergence of a ring-like structure. We will clarify the choice of windows and the dependence assessment in the next version. It’s a great suggestion to consider the role of rotating tuning curves in neural dynamics during interception.

      This paper shows sensory information can affect motor cortical activity whilst not affecting motor output. However, it is not the first to do so and fails to cite other papers that have investigated sensory modulation of the motor cortex (Stavinksy et al. 2017 Neuron, Pruszynski et al. 2011 Nature, Omrani et al. 2016 eLife). These studies should be mentioned in the Introduction to capture better the context around the present study. It would also be beneficial to add a discussion of how the results compare to the findings from these other works.

      Thanks for the reminder. We will introduce the relevant research in the next version of manuscript.

      This study also uses insights from single-unit analysis to inform mechanistic models of these population dynamics, which is a powerful approach, but is dependent on the validity of the single-cell analysis, which I have expanded on below.

      I have clarified some of the areas that would benefit from further analysis below:

      (1) Task:

      The task is well designed, although it would have benefited from perhaps one more target speed (for each direction). One monkey appears to have experienced one more target speed than the others (seen in Figure 3C). It would have been nice to have this data for all monkeys.

      Great suggestion! However, it’s hard to implement as the implanted arrays have been removed.

      (2) Single unit analyses:

      In some analyses, the effects of target speed look more driven by target movement direction (e.g. Figures 1D and E). To confirm target speed is the main modulator, it would be good to compare how much more variance is explained by models including speed rather than just direction. More target speeds may have been helpful here too.

      Nice suggestion! The fitting goodness of the simple model (just motor direction) is much less than the complex model (including target speed). We will update the results in the next version.

      The choice of the three categories (PD shift, gain addition) is not completely justified in a satisfactory way. It would be nice to see whether these three main categories are confirmed by unsupervised methods.

      A good point. We will have a try with unsupervised methods. 

      The decoder analyses in Figure 2 provide evidence that target speed modulation may change over the trial. Therefore, it is important to see how the window considered for the firing rate in Figure 1 (currently 100ms pre - 100ms post movement onset) affects the results.

      Thanks for the suggestion and close reading. We will test the decoder in other epochs.

      (3) Decoder:

      One feature of the task is that the reach endpoints tile the entire perimeter of the target circle (Figure 1B). However, this feature is not exploited for much of the single-unit analyses. This is most notable in Figure 2, where the use of a SVM limits the decoding to discrete values (the endpoints are divided into 8 categories). Using continuous decoding of hand kinematics would be more appropriate for this task.

      This is a very reasonable suggestion. In this study, we discrete the reach-direction as the previous studies (Li et al., 2018&2022) and thought that the discrete decoding was already enough to show the interaction of sensory and motor variables. In future studies, we will try continuous decoding of hand kinematics.

      (4) RNN:

      Mixed selectivity is not analysed in the RNN, which would help to compare the model to the real data where mixed selectivity is common. Furthermore, it would be informative to compare the neural data to the RNN activity using canonical correlation or Procrustes analyses. These would help validate the claim of similarity between RNN and neural dynamics, rather than allowing comparisons to be dominated by geometric similarities that may be features of the task. There is also an absence of alternate models to compare the perturbation model results to.

      Thank you for these helpful suggestions. We will perform decoding analysis on RNN units to verify if there is interaction of sensory and motor variables as in real data, as well as the canonical correlation or Procrustes analysis.

      Reviewer #2 (Public Review):

      Summary:

      In this manuscript, Zhang et al. examine neural activity in the motor cortex as monkeys make reaches in a novel target interception task. Zhang et al. begin by examining the single neuron tuning properties across different moving target conditions, finding several classes of neurons: those that shift their preferred direction, those that change their modulation gain, and those that shift their baseline firing rates. The authors go on to find an interesting, tilted ring structure of the neural population activity, depending on the target speed, and find that (1) the reach direction has consistent positioning around the ring, and (2) the tilt of the ring is highly predictive of the target movement speed. The authors then model the neural activity with a single neuron representational model and a recurrent neural network model, concluding that this population structure requires a mixture of the three types of single neurons described at the beginning of the manuscript.

      Strengths:

      I find the task the authors present here to be novel and exciting. It slots nicely into an overall trend to break away from a simple reach-to-static-target task to better characterize the breadth of how the motor cortex generates movements. I also appreciate the movement from single neuron characterization to population activity exploration, which generally serves to anchor the results and make them concrete. Further, the orbital ring structure of population activity is fascinating, and the modeling work at the end serves as a useful baseline control to see how it might arise.

      Thank you for recognizing our work.

      Weaknesses:

      While I find the behavioral task presented here to be excitingly novel, I find the presented analyses and results to be far less interesting than they could be. Key to this, I think, is that the authors are examining this task and related neural activity primarily with a single-neuron representational lens. This would be fine as an initial analysis since the population activity is of course composed of individual neurons, but the field seems to have largely moved towards a more abstract "computation through dynamics" framework that has, in the last several years, provided much more understanding of motor control than the representational framework has. As the manuscript stands now, I'm not entirely sure what interpretation to take away from the representational conclusions the authors made (i.e. the fact that the orbital population geometry arises from a mixture of different tuning types). As such, by the end of the manuscript, I'm not sure I understand any better how the motor cortex or its neural geometry might be contributing to the execution of this novel task.

      The present study shows the sensory modulation on motor tuning in single units and neural state during motor execution period. It’s a pity that the findings were constrained in certain time windows. We are still working this topic, and hopefully will address related questions in our follow-up studies.

      Main Comments:

      My main suggestions to the authors revolve around bringing in the computation through a dynamics framework to strengthen their population results. The authors cite the Vyas et al. review paper on the subject, so I believe they are aware of this framework. I have three suggestions for improving or adding to the population results:

      (1) Examination of delay period activity: one of the most interesting aspects of the task was the fact that the monkey had a random-length delay period before he could move to intercept the target. Presumably, the monkey had to prepare to intercept at any time between 400 and 800 ms, which means that there may be some interesting preparatory activity dynamics during this period. For example, after 400ms, does the preparatory activity rotate with the target such that once the go cue happens, the correct interception can be executed? There is some analysis of the delay period population activity in the supplement, but it doesn't quite get at the question of how the interception movement is prepared. This is perhaps the most interesting question that can be asked with this experiment, and it's one that I think may be quite novel for the field--it is a shame that it isn't discussed.

      Great idea! We are on the way, and close to complete the puzzle.

      (2) Supervised examination of population structure via potent and null spaces: simply examining the first three principal components revealed an orbital structure, with a seemingly conserved motor output space and a dimension orthogonal to it that relates to the visual input. However, the authors don't push this insight any further. One way to do that would be to find the "potent space" of motor cortical activity by regression to the arm movement and examine how the tilted rings look in that space (this is actually fairly easy to see in the reach direction components of the dPCA plot in the supplement--the rings will be highly aligned in this space). Presumably, then, the null space should contain information about the target movement. dPCA shows that there's not a single dimension that clearly delineates target speed, but the ring tilt is likely evident if the authors look at the highest variance neural dimension orthogonal to the potent space (the "null space")--this is akin to PC3 in the current figures, but it would be nice to see what comes out when you look in the data for it.

      Nice suggestion. Target-speed modulation mainly influences PC3, which is consistent with ‘null space’ hypothesis. We will try other methods of dimensionality reduction (e.g. dPCA, Manopt) to determine the potent and null space.

      (3) RNN perturbations: as it's currently written, the RNN modeling has promise, but the perturbations performed don't provide me with much insight. I think this is because the authors are trying to use the RNN to interpret the single neuron tuning, but it's unclear to me what was learned from perturbing the connectivity between what seems to me almost arbitrary groups of neurons (especially considering that 43% of nodes were unclassifiable). It seems to me that a better perturbation might be to move the neural state before the movement onset to see how it changes the output. For example, the authors could move the neural state from one tilted ring to another to see if the virtual hand then reaches a completely different (yet predictable) target. Moreover, if the authors can more clearly characterize the preparatory movement, perhaps perturbations in the delay period would provide even more insight into how the interception might be prepared.

      We are sorry that we didn’t clarify the definition of “none” type, which can be misleading. The 43% unclassified nodes include those inactive ones, when only activate (task-related) nodes included, the ratio of unclassified nodes would be much lower. By perturbing the connectivity, we intended to explore the interaction between different modulations.

      Thank you for the great advice. We tried moving neural states from one ring to another without changing the directional cluster, but this perturbation didn’t have a significant influence on network performance as expected. We will check this result again and try perturbations in the delay period.

      Reviewer #3 (Public Review):

      Summary:

      This experimental study investigates the influence of sensory information on neural population activity in M1 during a delayed reaching task. In the experiment, monkeys are trained to perform a delayed interception reach task, in which the goal is to intercept a potentially moving target.

      This paradigm allows the authors to investigate how, given a fixed reach endpoint (which is assumed to correspond to a fixed motor output), the sensory information regarding the target motion is encoded in neural activity.

      At the level of single neurons, the authors found that target motion modulates the activity in three main ways: gain modulation (scaling of the neural activity depending on the target direction), shift (shift of the preferred direction of neurons tuned to reach direction), or addition (offset to the neural activity).

      At the level of the neural population, target motion information was largely encoded along the 3rd PC of the neural activity, leading to a tilt of the manifold along which reach direction was encoded that was proportional to the target speed. The tilt of the neural manifold was found to be largely driven by the variation of activity of the population of gain-modulated neurons.

      Finally, the authors studied the behaviour of an RNN trained to generate the correct hand velocity given the sensory input and reach direction. The RNN units were found to similarly exhibit mixed selectivity to the sensory information, and the geometry of the « neural population » resembled that observed in the monkeys.

      Strengths:

      - The experiment is well set up to address the question of how sensory information that is directly relevant to the behaviour but does not lead to a direct change in behavioural output modulates motor cortical activity.

      - The finding that sensory information modulates the neural activity in M1 during motor preparation and execution is non trivial, given that this modulation of the activity must occur in the nullspace of the movement.

      - The paper gives a complete picture of the effect of the target motion on neural activity, by including analyses at the single neuron level as well as at the population level. Additionally, the authors link those two levels of representation by highlighting how gain modulation contributes to shaping the population representation.

      Thanks for your recognition.

      Weaknesses:

      - One of the main premises of the paper is the fact that the motor output for a given reach point is preserved across different target motions. However, as the authors briefly mention in the conclusion, they did not record muscle activity during the task, but only hand velocity, making it impossible to directly verify how preserved muscle patterns were across movements. While the authors highlight that they did not see any difference in their results when resampling the data to control for similar hand velocities across conditions, this seems like an important potential caveat of the paper whose implications should be discussed further or highlighted earlier in the paper.

      Thanks for the suggestion. We will highlight the resampling results as important control in the next version of manuscript.

      - The main takeaway of the RNN analysis is not fully clear. The authors find that an RNN trained given a sensory input representing a moving target displays modulation to target motion that resembles what is seen in real data. This is interesting, but the authors do not dissect why this representation arises, and how robust it is to various task design choices. For instance, it appears that the network should be able to solve the task using only the motion intention input, which contains the reach endpoint information. If the target motion input is not used for the task, it is not obvious why the RNN units would be modulated by this input (especially as this modulation must lie in the nullspace of the movement hand velocity if the velocity depends only on the reach endpoint). It would thus be important to see alternative models compared to true neural activity, in addition to the model currently included in the paper. Besides, for the model in the paper, it would therefore be interesting to study further how the details of the network setup (eg initial spectral radius of the connectivity, weight regularization, or using only the target position input) affect the modulation by the motion input, as well as the trained population geometry and the relative ratios of modulated cells after training.

      Great suggestions. It’s a considerable pity that we didn’t dissect the formation reason and influence factor of the representation in the current version. We’ve tried several combinations of inputs before: in the network which received only motor intention and GO inputs, there were rings but not tilting related to target-speed; in the network which received only target location and GO inputs, there were ring-like structures but not clear directional clusters. We will check these results and try alternative models in the next version. In future studies, we will examine the influence of network setup details.

      - Additionally, it is unclear what insights are gained from the perturbations to the network connectivity the authors perform, as it is generally expected that modulating the connectivity will degrade task performance and the geometry of the responses. If the authors wish the make claims about the role of the subpopulations, it could be interesting to test whether similar connectivity patterns develop in networks that are not initialized with an all-to-all random connectivity or to use ablation experiments to investigate whether the presence of multiple types of modulations confers any sort of robustness to the network.

      Thank you for the great suggestions. By perturbations, we intended to explore the contribution of interaction between certain subpopulations. We tried ablation experiments, but the result was not significant. Probably because the most units were of mixed selectivity, the units of only modulations were not enough for bootstrapping, or the random sampling from single subpopulation (bearing mixed selectivity) could be repeated. We will consider these suggestions carefully in the revised version.

      - The results suggest that the observed changes in motor cortical activity with target velocity result from M1 activity receiving an input that encodes the velocity information. This also appears to be the assumption in the RNN model. However, even though the input shown to the animal during preparation is indeed a continuously moving target, it appears that the only relevant quantity to the actual movement is the final endpoint of the reach. While this would have to be a function of the target velocity, one could imagine that the computation of where the monkeys should reach might be performed upstream of the motor cortex, in which case the actual target velocity would become irrelevant to the final motor output. This makes the results of the paper very interesting, but it would be nice if the authors could discuss further when one might expect to see modulation by sensory information that does not directly affect motor output in M1, and where those inputs may come from. It may also be interesting to discuss how the findings relate to previous work that has found behaviourally irrelevant information is being filtered out from M1 (for instance, Russo et al, Neuron 2020 found that in monkeys performing a cycling task, context can be decoded from SMA but not from M1, and Wang et al, Nature Communications 2019 found that perceptual information could not be decoded from PMd)?

      How and where sensory information modulates M1 are very interesting and open questions. We will discuss further about this topic in the next version.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      Semenova et al. have studied a large cross-sectional cohort of people living with HIV on suppressive ART, N=115, and performed high dimensional flow cytometry to then search for associations between immunological and clinical parameters and intact/total HIV DNA levels.

      A number of interesting data science/ML approaches were explored on the data and the project seems a serious undertaking. However, like many other studies that have looked for these kinds of associations, there was not a very strong signal. Of course, the goal of unsupervised learning is to find new hypotheses that aren't obvious to human eyes, but I felt in that context, there were (1) results slightly oversold, (2) some questions about methodology in terms mostly of reservoir levels, and (3) results were not sufficiently translated back into meaning in terms of clinical outcomes.

      We appreciate the reviewer’s perspective.  In our revised version of the manuscript, we have attempted to address these concerns by more adequately explaining the limitations of the study and by more thoroughly discussing the context of the findings.  We are not able to associate the findings with specific clinical outcomes for individual study participants but we speculate about the overall biological meaning of these associations across the cohort.  We cannot disagree with the reviewer, but we find the associations statistically significant, potentially reflecting real biological associations, and forming the basis for future hypothesis testing research. 

      Strengths:

      The study is evidently a large and impressive undertaking and combines many cutting-edge statistical techniques with a comprehensive experimental cohort of people living with HIV, notably inclusive of populations underrepresented in HIV science. A number of intriguing hypotheses are put forward that could be explored further. Sharing the data could create a useful repository for more specific analyses.

      We thank the reviewer for this assessment.

      Weaknesses:

      Despite the detailed experiments and methods, there was not a very strong signal for the variable(s) predicting HIV reservoir size. The Spearman coefficients are ~0.3, (somewhat weak, and acknowledged as such) and predictive models reach 70-80% prediction levels, though sometimes categorical variables are challenging to interpret.

      We agree with the reviewer that individual parameters are only weakly correlated with the HIV reservoir, likely reflecting the complex and multi-factorial nature of reservoir/immune cell interactions.  Nevertheless, these associations are statistically significant and form the basis for functional testing in viral persistence.

      There are some questions about methodology, as well as some conclusions that are not completely supported by results, or at minimum not sufficiently contextualized in terms of clinical significance.  On associations: the false discovery rate correction was set at 5%, but data appear underdetermined with fewer observations than variables (144vars > 115ppts), and it isn't always clear if/when variables are related (e.g inverses of one another, for instance, %CD4 and %CD8).

      When deriving a list of cell populations whose frequency would be correlated with the reservoir, we focused on well-defined cell types for which functional validation exists in the literature to consider them as distinct cell types.  For many of the populations, gating based on combinations of multiple markers leads to recovery of very few cells, and so we excluded some potential combinations from the analysis.  We are also making our raw data available for others to examine and find associations not considered by our manuscript.

      The modeling of reservoir size was unusual, typically intact and defective HIV DNA are analyzed on a log10 scale (both for decays and predicting rebound). Also, sometimes in this analysis levels are normalized (presumably to max/min?, e.g. S5), and given the large within-host variation of level we see in other works, it is not trivial to predict any downstream impact of normalization across population vs within-person.

      We have repeated the analysis using log10 transformed data and the new figures are shown in Figure 1 and S2-S5.

      Also, the qualitative characterization of low/high reservoir is not standard and naturally will split by early/later ART if done as above/below median. Given the continuous nature of these data, it seems throughout that predicting above/below median is a little hard to translate into clinical meaning.

      Our ML models included time before ART as a variable in the analysis, and this was not found to be a significant driver of the reservoir size associations, except for the percentage of intact proviruses (see Figure 2C). Furthermore, we analyzed whether any of the reservoir correlated immune variables were associated with time on ART and found that, although some immune variables are associated with time on therapy, this was not the case for most of them (Table S4). We agree that it is challenging to translate above or below median into clinical meaning for this cohort, but we emphasize that this study is primarily a hypothesis generating approach requiring additional validation for the associations observed.  We attempted to predict reservoir size as a continuous variable using the data and this approach was not successful (Figure S13). We believe that a significantly larger cohort will likely be required to generate a ML model that can accurately predict the reservoir as a continuous variable.  We have added additional discussion of this to the manuscript.

      Lastly, the work is comprehensive and appears solid, but the code was not shared to see how calculations were performed.

      We now provide a link to the code used to perform the analyses in the manuscript, https://github.com/lesiasemenova/ML_HIV_reservoir.

      Reviewer #2 (Public Review):

      Summary:

      Semenova et. al., performed a cross-sectional analysis of host immunophenotypes (using flow cytometry) and the peripheral CD4+ T cell HIV reservoir size (using the Intact Proviral DNA Assay, IPDA) from 115 people with HIV (PWH) on ART. The study mostly highlights the machine learning methods applied to these host and viral reservoir datasets but fails to interpret these complex analyses into (clinically, biologically) interpretable findings. For these reasons, the direct translational take-home message from this work is lost amidst a large list of findings (shown as clusters of associated markers) and sentences such as "this study highlights the utility of machine learning approaches to identify otherwise imperceptible global patterns" - lead to overinterpretation of their data.

      We have addressed the reviewer’s concern by modifications to the manuscript that enhance the interpretation of the findings in a clinical and biological context.

      Strengths:

      Measurement of host immunophenotyping measures (multiparameter flow cytometry) and peripheral HIV reservoir size (IPDA) from 115 PWH on ART.

      Major Weaknesses:

      (1) Overall, there is little to no interpretability of their machine learning analyses; findings appear as a "laundry list" of parameters with no interpretation of the estimated effect size and directionality of the observed associations. For example, Figure 2 might actually give an interpretation of each X increase in immunophenotyping parameter, we saw a Y increase/decrease in HIV reservoir measure.

      We have added additional text to the manuscript in which we attempt to provide more immunological and clinical interpretation of the associations.  We also have emphasized that these associations are still speculative and will require additional validation.  Nevertheless, our data should provide a rich source of new hypotheses regarding immune system/reservoir interaction that could be tested in future work.

      (2) The correlations all appear to be relatively weak, with most Spearman R in the 0.30 range or so.

      We agree with the review that the associations are mostly weak, consistent with previous studies in this area.  This likely is an inherent feature of the underlying biology – the reservoir is likely associated with the immune system in complex ways and involves stochastic processes that will limit the predictability of reservoir size using any single immune parameter. We have added additional text to the manuscript to make this point clearer.

      (3) The Discussion needs further work to help guide the reader. The sentence: "The correlative results from this present study corroborate many of these studies, and provide additional insights" is broad. The authors should spend some time here to clearly describe the prior literature (e.g., describe the strength and direction of the association observed in prior work linking PD-1 and HIV reservoir size, as well as specify which type of HIV reservoir measures were analyzed in these earlier studies, etc.) and how the current findings add to or are in contrast to those prior findings.

      We have added additional text to the manuscript to help guide the readers through the possible biological significance of the findings and the context with respect to prior literature.

      (4) The most interesting finding is buried on page 12 in the Discussion: "Uniquely, however, CD127 expression on CD4 T cells was significantly inversely associated with intact reservoir frequency." The authors should highlight this in the abstract, and title, and move this up in the Discussion. The paper describes a very high dimensional analysis and the key takeaways are not clear; the more the author can point the reader to the take-home points, the better their findings can have translatability to future follow-up mechanistic and/or validation studies.

      We appreciate the reviewer’s comment.  We have increased the emphasis on this finding in the revised version of the manuscript.

      (5) The authors should avoid overinterpretation of these results. For example in the Discussion on page 13 "The existence of two distinct clusters of PWH with different immune features and reservoir characteristics could have important implications for HIV cure strategies - these two groups may respond differently to a given approach, and cluster membership may need to be considered to optimize a given strategy." It is highly unlikely that future studies will be performing the breadth of parameters resulting here and then use these directly for optimizing therapy.

      Our analyses indicate that membership of study participants in cluster1 or cluster 2 can be fairly accurately determined by a small number of individual parameters (KLRG1 etc, Figure 4F), and measuring the cells of PWH with the degree of breadth used in this paper would not be necessary to classify PWH into these clusters.  As such, we feel that it is not unrealistic to speculate that this finding could turn out to be clinically useful, if it becomes clear that the clusters are biologically meaningful.

      (6) There are only TWO limitations listed here: cross-sectional study design and the use of peripheral blood samples. (The subsequent paragraph notes an additional weakness which is misclassification of intact sequences by IPDA). This is a very limited discussion and highlights the need to more critically evaluate their study for potential weaknesses.

      We have expanded on the list of limitations discussed in the manuscript. In particular, we now address the size of the cohort, the composition with respect to different genders and demographics, lack of information for the timing of ART and the lack of information regarding intracellular transcriptional pathways.

      (7) A major clinical predictor of HIV reservoir size and decay is the timing of ART initiation. The authors should include these (as well as other clinical covariate data - see #12 below) in their analyses and/or describe as limitations of their study.

      All of the participants that make up our cohort were treated during chronic infection, and the precise timing of ART initiation is unclear in most of these cases.  We have added additional information to explain this in the manuscript and include this in the list of limitations.

      Reviewer #3 (Public Review):

      Summary:

      This valuable study by Semenova and colleagues describes a large cross-sectional cohort of 115 individuals on ART. Participants contributed a single blood sample which underwent IPDA, and 25-color flow with various markers (pre and post-stimulation). The authors then used clustering, decision tree analyses, and machine learning to look for correlations between these immunophenotypic markers and several measures of HIV reservoir volume. They identified two distinct clusters that can be somewhat differentiated based on total HIV DNA level, intact HIV DNA level, and multiple T cell cellular markers of activation and exhaustion.

      The conclusions of the paper are supported by the data but the relationships between independent and dependent variables in the models are correlative with no mechanistic work to determine causality. It is unclear in most cases whether confounding variables could explain these correlations. If there is causality, then the data is not sufficient to infer directionality (ie does the immune environment impact the HIV reservoir or vice versa or both?). In addition, even with sophisticated and appropriate machine learning approaches, the models are not terribly predictive or highly correlated. For these reasons, the study is very much hypothesis-generating and will not impact cure strategies or HIV reservoir measurement strategies in the short term.

      We appreciate the reviewer’s comments regarding the value of our study.  We fully acknowledge that the causal nature and directionality of these associations are not yet clear and agree that the study is primarily hypothesis generating in nature.  Nevertheless, we feel that the hypotheses generated will be valuable to the field.  We have added additional text to the manuscript to emphasize the hypothesis generating nature of this paper.

      Strengths:

      The study cohort is large and diverse in terms of key input variables such as age, gender, and duration of ART. Selection of immune assays is appropriate. The authors used a wide array of bioinformatic approaches to examine correlations in the data. The paper was generally well-written and appropriately referenced.

      Weaknesses:

      (1) The major limitation of this work is that it is highly exploratory and not hypothesis-driven. While some interesting correlations are identified, these are clearly hypothesis-generating based on the observational study design.

      We agree that the major goal of this study was hypothesis generating and that our work is exploratory in nature. Performing experiments with mechanism testing goals in human participants with HIV is challenging.  Additionally, before such mechanistic studies can be undertaken, one must have hypotheses to test. As such we feel our study will be useful for the field in helping to identify hypotheses that could potentially be tested.

      (2) The study's cross-sectional nature limits the ability to make mechanistic inferences about reservoir persistence. For instance, it would be very interesting to know whether the reservoir cluster is a feature of an individual throughout ART, or whether this outcome is dynamic over time.

      We agree with the reviewer’s comment. Longitudinal studies are challenging to carry out with a study cohort of this size, and addressing questions such as the one raised by the reviewer would be of great interest. We believe our study nevertheless has value in identifying hypotheses that could be tested in a longitudinal study.

      (3) A fundamental issue is that I am concerned that binarizing the 3 reservoir metrics in a 50/50 fashion is for statistical convenience. First, by converting a continuous outcome into a simple binary outcome, the authors lose significant amounts of quantitative information. Second, the low and high reservoir outcomes are not actually demonstrated to be clinically meaningful: I presume that both contain many (?all) data points above levels where rebound would be expected soon after interruption of ART. Reservoir levels would also have no apparent outcome on the selection of cure approaches. Overall, dividing at the median seems biologically arbitrary to me.

      The reviewer raises a valid point that the clinical significance of above or below median reservoir metrics is unclear, and that the size of the reservoir has potentially little relation to rebound and cure approaches.  In the manuscript, we attempted to generate models that can predict reservoir size as a continuous variable in Figure S13 and find that this approach performs poorly, while a binarized approach was more successful. As such we have included both approaches in the manuscript.  It is possible that future studies with larger sample sizes and more detailed measurements will perform better for continuous variable prediction.  While this is a fairly large study (n=115) by the standards of HIV reservoir analyses, it is a small study by the standards of the machine learning field, and accurate predictive ML models for reservoir size as a continuous variable will likely require a much larger set of samples/participants.  Nevertheless, we feel our work has value as a template for ML approaches that may be informative for understanding HIV/immune interactions and generates novel hypotheses that could be validated by subsequent studies.

      (4) The two reservoir clusters are of potential interest as high total and intact with low % intact are discriminated somewhat by immune activation and exhaustion. This was the most interesting finding to me, but it is difficult to know whether this clustering is due to age, time on ART, other co-morbidity, ART adherence, or other possible unmeasured confounding variables.

      We agree that this finding is one of the more interesting outcomes of the study. We examined a number of these variables for association with cluster membership, and these data are reported in Figure S8A-D.  Age, years of ART and CD4 Nadir were all clearly different between the clusters.   The striking feature of this clustering, however, is the clear separation between the two groups of participants, as opposed to a continuous gradient of phenotypes.  This could reflect a bifurcation of outcomes for people with HIV, dynamic changes in the reservoir immune interactions over time, or different levels of untreated infection.  It is certainly possible that some other unmeasured confounding variables contribute to this outcome and we have attempted to make this limitation clearer.

      (5) At the individual level, there is substantial overlap between clusters according to total, intact, and % intact between the clusters. Therefore, the claim in the discussion that these 2 cluster phenotypes may require different therapeutic approaches seems rather speculative. That said, the discussion is very thoughtful about how these 2 clusters may develop with consideration of the initial insult of untreated infection and / or differences in immune recovery.

      We agree with the reviewer that this claim is speculative, and we have attempted to moderate the language of the text in the revised version.

      (6) The authors state that the machine learning algorithms allow for reasonable prediction of reservoir volume. It is subjective, but to me, 70% accuracy is very low. This is not a disappointing finding per se. The authors did their best with the available data. It is informative that the machine learning algorithms cannot reliably discriminate reservoir volume despite substantial amounts of input data. This implies that either key explanatory variables were not included in the models (such as viral genotype, host immune phenotype, and comorbidities) or that the outcome for testing the models is not meaningful (which may be possible with an arbitrary 50/50 split in the data relative to median HIV DNA volumes: see above).

      We acknowledge that the predictive power of the models generated from these data is modest and we have clarified this point in the revised manuscript. As the reviewer indicates, this may result from the influence of unmeasured variables and possible stochastic processes.  The data may thus demonstrate a limit to the predictability of reservoir size which may be inherent to the underlying biology.  As we mention above, this study size (n-115) is fairly small for the application of ML methods, and an increased sample size will likely improve the accuracy of the models. At this stage, the models we describe are not yet useful as predictive clinical tools, but are still nonetheless useful as tools to describe the structure of the data and identify reservoir associated immune cell types.

      (7) The decision tree is innovative and a useful addition, but does not provide enough discriminatory information to imply causality, mechanism, or directionality in terms of whether the immune phenotype is impacting the reservoir or vice versa or both. Tree accuracy of 80% is marginal for a decision tool.

      The reviewer is correct about these points.  In the revised manuscript, we have attempted to make it clear that we are not yet advocating using this approach as a decision tool, but simply a way to visualize the data and understand the structure of the dataset.  As we discuss above, the models will likely need to be trained on a larger dataset and achieve higher accuracy before use as a decision tool.

      (8) Figure 2: this is not a weakness of the analysis but I have a question about interpretation. If total HIV DNA is more predictive of immune phenotype than intact HIV DNA, does this potentially implicate a prior high burden of viral replication (high viral load &/or more prolonged time off ART) rather than ongoing reservoir stimulation as a contributor to immune phenotype? A similar thought could be applied to the fact that clustering could only be detected when applied to total HIV DNA-associated features. Many investigators do not consider defective HIV DNA to be "part of the reservoir" so it is interesting to speculate why these defective viruses appear to have more correlation with immunophenotype than intact viruses.

      We agree with the reviewer that this observation could reflect prior viral burden and we have added additional text to make this clearer.  Even so, we cannot rule out a model in which defective viral DNA is engaged in ongoing stimulation of the immune system during ART, leading to the stronger association between total DNA and the immune cell phenotypes. We hypothesize that the defective proviruses could potentially be triggering innate immune pattern recognition receptors via viral RNA or DNA, and a higher burden of the total reservoir leads to a stronger apparent association with the immune phenotype.  We have included text in the discussion about this hypothesis.

      (9) Overall, the authors need to do an even more careful job of emphasizing that these are all just correlations. For instance, HIV DNA cannot be proven to have a causal effect on the immunophenotype of the host with this study design. Similarly, immunophenotype may be affecting HIV DNA or the correlations between the two variables could be entirely due to a separate confounding variable

      We have revised the text of the manuscript to emphasize this point, and we acknowledge that any causal relationships are, at this point, simply speculation. 

      (10) In general, in the intro, when the authors refer to the immune system, they do not consistently differentiate whether they are referring to the anti-HIV immune response, the reservoir itself, or both. More specifically, the sentence in the introduction listing various causes of immune activation should have citations. (To my knowledge, there is no study to date that definitively links proviral expression from reservoir cells in vivo to immune activation as it is next to impossible to remove the confounding possible imprint of previous HIV replication.) Similarly, it is worth mentioning that the depletion of intact proviruses is quite slow such that provial expression can only be stimulating the immune system at a low level. Similarly, the statement "Viral protein expression during therapy likely maintains antigen-specific cells of the adaptive immune system" seems hard to dissociate from the persistence of immune cells that were reactive to viremia.

      We updated the text of the manuscript to address these points and have added additional citations as per the reviewer’s suggestion.

      (11) Given the many limitations of the study design and the inability of the models to discriminate reservoir volume and phenotype, the limitations section of the discussion seems rather brief.

      We have now expanded the limitations section of the discussion and added additional considerations. We now include a discussion of the study cohort size, composition and the detail provided by the assays.

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors):

      A few specific comments:

      "This pattern is likely indicative of a more profound association of total HIV DNA with host immunophenotype relative to intact HIV DNA."

      Most studies I have seen (e.g. single cell from Lictherfeld/Yu group) show intact proviruses are generally more activated/detectable/susceptible to immune selection, so I have a hard time thinking defective proviruses are actually more affected by immunotype.

      We hypothesize that this association is actually occurring in the opposite direction – that the defective provirus are having a greater impact on the immune phenotype, due to their greater number and potential ability to engage innate or adaptive immune receptors. We have clarified this point in the manuscript

      "The existence of two distinct clusters of PWH with different immune features and reservoir characteristics could have important implications for HIV cure strategies - these two groups may respond differently to a given approach, and cluster membership may need to be considered to optimize a given strategy."

      I find this a bit of a reach, given that the definition of 2 categories depended on the total size.

      We have modified the language of this section to reduce the level of speculation.

      "This study is cross-sectional in nature and is primarily observational, so caution should be used interpreting findings associated with time on therapy".

      I found this an interesting statement because ultimately time on ART shows up throughout the analysis as a significant predictor, do you mean something about how time on ART could indicate other confounding variables like ART regimen or something?

      We have rephrased this comment to avoid confusion.  We were simply trying to make the point that we should avoid speculating about longitudinal dynamics from cross sectional data.

      "As expected, the plots showed no significant correlation for intact HIV DNA versus years of ART (Figure 1B), while total reservoir size was positively correlated with the time of ART (Figure 1A, Spearman r = 0.31)."<br />  Is this expected? Studies with longitudinal data almost uniformly show intact decay, at least for the first 10 or so years of ART, and defective/total stability (or slight decay). Also probably "time on ART" to not confuse with the duration of infection before ART.

      We have updated the language of this section to address this comment.  We have avoided comparing our data with respect to time on ART to longitudinal studies for reasons given above.

      On dimensionality reduction, as this PaCMAP seems a relatively new technique (vs tSNE and UMAP which are more standard, but absolutely have their weaknesses), it does seem important to contextualize. I think it would still be useful to show PCA and asses the % variance of each additional dimension to assess the effective dimensionality, it would be helpful to show a plot of % variance by # components to see if there is a cutoff somewhere, and if PaCMAP is really picking this up to determine the 2 dimensions/2 clusters is ideal. Figure 4B ultimately shows a lot of low/high across those clusters, and since low/high is defined categorically it's hard to know which of those dots are very close to the other categories.

      We have added this analysis to the manuscript – found in Figure S9. The PCA plot indicates that members of the two clusters also separate on PCA although this separation is not as clear as for the PaCMAP plot.

      Minor comments on writing etc:

      Intro

      -Needs some references on immune activation sequelae paragraph.

      We have added some additional references to this section.

      -"promote the entry of recently infected cells into the reservoir" -- that is only one possible mechanistic explanation, it's not unreasonable but it seems important to keep options open until we have more precise data that can illuminate the mechanism of the overabundance.

      We have modified the text to discuss additional hypotheses.

      -You might also reference Pankau et al Ppath for viral seeding near the time of ART.

      We have added this reference.

      -"Viral protein expression during therapy likely maintains antigen-specific cells of the adaptive immune system" - this was unclear to me, do you mean HIV-specific cells that act against HIV during ART? I think most studies show immunity against HIV (CD8 and CD4) wanes over time during ART.

      The Goonetilleke lab has recently generated data indicating that antiviral T cell responses are remarkably stable over time on ART, but we agree with the reviewer that the idea that ongoing antigen expression in the reservoir maintains these cells is speculative.  We have modified the text to make this point clearer.

      -Overall I think the introduction lacked a little bit of definitional precision: i.e. is the reservoir intact vs replication competent vs all HIV DNA and whether we are talking about PWH on long-term ART and how long we should be imagining? The first years of ART are certainly different than later, in terms of dynamics. The ultimate implications are likely specific for some of these categorizations.

      -"persistent sequelae of the massive disruptions to T cell homeostasis and lymphoid structures that occur during untreated HIV infection" needs a lot more context/referencing. For instance, Peter Hunt showed a decrease in activation after ART a long time ago.

      -Heather Best et al show T cell clonality stays perturbed after ART.

      We have updated the text of the introduction and added references to address the reviewer’s comments.

      Results

      -It would be important to mention the race of participants and any information about expected clades of acquired viruses, this gets mentioned eventually with reference to the Table but the breakdown would be helpful right away.

      We have added this information to the results section.

      -"performed Spearman correlations", may be calculated or tested?

      We have corrected the language for this sentence.

      Comments on figures:

      -Figure 1 data on linear scale (re discussion above) -- hard to even tell if there is a decay (to match with all we know from various long-term ART studies).

      -Figure 4 data is shown on ln (log_e) scale, which is hard to interpret for most people.

      -Figures 4 C,D, and E should have box plots to visually assess the significance.

      -Figure 4B legend says purple/pink but I think the colors are different in the plot, could be about transparency

      -Figure 5 it is now not clear if log_e(?).

      -Figure 6 "HIV reservoir characteristics" might be better to make this more explicit. Do you mean for instance in the 6B title Total HIV DNA per million CD4+ T cells I think?

      We have made these modifications.

      Reviewer #2 (Recommendations For The Authors):

      Minor Weaknesses:

      (1) The Introduction is too long and much of the text is not directly related to the study's research question and design.

      We have streamlined the introduction in the revised manuscript.

      (2) While no differences were seen by age or race, according to the authors, this is unlikely to be useful since the numbers are so small in some of these subcategories. Results from sensitivity analyses (e.g., excluding these individuals) may be more informative/useful.

      We agree that the lower numbers of participants for some subgroupings makes it challenging to know for sure if there are any differences based on these variables.  Have added text to clarify this. We have added age, race and gender to the LOCO analysis and to the variable inflation importance analysis (Table S5).

      (3) For Figure 4, based on what was described in the Results section of the manuscript, the authors should clarify that the figures show results for TOTAL HIV DNA only (not intact DNA): "Dimension reduction machine learning approaches identified two robust clusters of PWH when using total HIV DNA reservoir-associated immune cell frequencies (Figure 4A), but not for intact or percentage intact HIV DNA (Figure 4B and 4C)".

      We have added this information.

      (4) The statement on page 5, first paragraph, "Interestingly, when we examined a plot of percent intact proviruses versus time on therapy (Figure 1C), we observed a biphasic decay pattern," is not new (Peluso JCI Insight 2020, Gandhi JID 2023, McMyn JCI 2023). Prior studies have clearly demonstrated this biphasic pattern and should be cited here, and the sentence should be reworded with something like "consistent with prior work", etc.

      We have added citations to these studies and rephrased this comment.

      (5) The Cohort and sample collection sections are somewhat thin. Further details on the cohort details should include at the very minimum some description of the timing of ART initiation (is this mostly a chronic-treated cohort?) and important covariate data such as nadir CD4+ T cell count, pre-ART viral load, duration of ART suppression, etc.

      The cohort was treated during chronic infection, and we have clarified this in the manuscript.  Information regarding CD4 nadir and years on ART are included in Table 1.  Unfortunately, pre-ART viral load was not available for most members of this cohort, so we did not use it for analyses. The partial pre-ART viral load data is included with the dataset we are making publicly available.

      Reviewer #3 (Recommendations For The Authors):

      Minor points:

      (1) What is meant by CD4 nadir? Is this during primary infection or the time before ART initiation?

      We have clarified this description in the manuscript.  This term refers to the lowest CD4 count recorded during untreated infection.

      (2) The authors claim that determinants of reservoir size are starting to emerge but other than the timing of ART, I am not sure what studies they are referring to.

      We have updated the language of this section.  We intended to refer to studies looking at correlates of reservoir size, and feel that this is a more appropriate term that ‘determinants’

      (3) The discussion does not tie in the model-generated hypotheses with the known mechanisms that sustain the reservoir: clonal proliferation balanced by death and subset differentiation. It would be interesting to tie in the proposed reservoir clusters with these known mechanisms.

      We have added additional text to the manuscript to address these mechanisms.

      (4) Figure 1: Total should be listed as total HIV DNA.

      We have updated this in the manuscript.

      (5) Figure 1C: Worth mentioning the paper by Reeves et al which raises the possibility that the flattening of intact HIV DNA at 9 years may be spurious due to small levels of misclassification of defective as intact.

      We have added this reference.

      (6) "Total reservoir frequency" should be "total HIV DNA concentration"

      We respectfully feel that “frequency” is a more accurate term than “concentration”, since we are expressing the reservoir as a fraction of the CD4 T cells, while “concentration” suggests a denominator of volume.

      (7) Figure S2-5: label y-axis total HIV DNA.

      We have updated this figure.

    1. Note: This response was posted by the corresponding author to Review Commons. The content has not been altered except for formatting.

      Learn more at Review Commons


      Reply to the reviewers

      Rebuttal_ Preprint- #RC-2023-02144

      First of all we would like to thank the three reviewers for their constructive and positive comments and suggestions, and the time spent in reviewing our manuscript. Their suggestions and comments had contributed to improve our manuscript. We feel the manuscript is much strengthened by this revision.

      Reviewer #1 (Evidence, reproducibility and clarity (Required)):

      __Summary:____ __The manuscript by Dabsan et al builds on earlier work of the Igbaria lab, who showed that ER-luminal chaperones can be refluxed into the cytosol (ERCYS) during ER stress, which constitutes a pro-survival pathway potentially used by cancer cells. In the current work, they extent these observations and a role for DNAJB12&14 in ERCYS. The work is interesting and the topic is novel and of great relevance for the proteostasis community. I have a number of technical comments:

      We thank the reviewer for his/her positive comments on our manuscript.


      __Major and minor comments: __

      1- In the description of Figure 2, statistics is only show to compare untreated condition with those treated with Tg or Tm, but no comparison between condition and different proteins. As such, the statement made by the authors "...DNAJB14-silenced cells were only affected in AGR2 but not in DNAJB11 or HYOU1 cytosolic accumulation" cannot be made.

      Answer: We totally agree with the reviewer#1. The aim of this figure is to show that during ER stress, a subset of ER proteins are refluxed to the cytosol. This is happening in cells expressing DNAJB12 and DNAJB14. We are not comparing the identity of the expelled proteins between DNAJB12-KD cells and DNAJB14-KD cells, This is not the scoop of this paper as such the statement was removed.

      2- Figure S2C: D11 seems to increase in the cytosolic fraction after Tm and Tg treatment. However, this is not reflected in the text. The membrane fraction also increases in the DKO. Is the increase of D11 in both cytosol and membrane and indication for a transcriptional induction of this protein by Tm/Tg? Again, the authors are not reflecting on this in their text.

      Answer: We performed qPCR experiments in control, DNAJB12-KD, DNAJB14-KD and in the DNAJB12/DNAJB14 double knock down cells (in both A549 and PC3 cells) to follow the mRNA levels of DNAJB11. As shown in (Figure S2F-S2N), there is no increase in the mRNA levels of DNAJB11, AGR2 or HYOU1 in the different cells in normal (unstressed conditions). Upon ER stress with tunicamycin or thapsigargin there is a little increase in the mRNA levels of HYOU1 and AGR2 but not in DNAJB11 mRNA levels. On the other hand, we also performed western blot analysis and we did not detect any difference between the different knockdown cells when we analyzed the levels of DNAJB11 compared to GAPDH. Those data are now added as (Figure S2F-S2N).

      We must note that although AGR2 and HYOU1 are induced at the mRNA as a result of ER stress, the data with the overexpression of DNAJB12 and DNAJB14 are important as control experiments because when DNAJB12 is overexpressed it doesn’t inducing the ER stress (Figure S3C-S3D). In those conditions there is an increase of the cytosolic accumulation of AGR2, HYOU1 and DNAJB11 despite that there was no induction of AGR2, HYOU1 or DNAJB11 (Figure 3C and Figure 3E, Figure S3, Figure 4, and Figure S4) . Those results argue against the idea that the reflux is a result of protein induction and an increase in the total proteins levels.

      3- Figure 2D: Only p21 is quantified. phospho-p53 and p53 levels are not quantified.


      Answer: We added the quantification of phospho-p53 and the p53 levels to (Figure 2E-G). Additional blots of the P21, phosphor-p53 and p53 now added to FigureS2O.

      4- Figure 2D: There appears to be a labelling error

      Answer: Yes, the labelling error was corrected.

      5- Are there conditions where DNAJB12 would be higher?

      Answer: In some cancer types there is a higher DNAJB12, DNAJB14 and SGTA expression levels that are associated with poor prognosis and reduced survival (New Figure S6E-M). The following were added to the manuscript: “Finally, we tested the effect of DNAJB12, DNAJB14, and SGTA expression levels on the survival of cancer patients. A high copy number of DNAJB12 is an unfavorable marker in colorectal cancer and in head and neck cancer because it is associated with poor prognosis in those patients (Figure S6E). A high copy number of DNAJB12, DNAJB14, and SGTA is associated with poor prognosis in many other cancer types, including colon adenocarcinoma (COAD), acute myeloid leukemia (LAML), adrenocortical carcinoma (ACC), mesothelioma (MESO), and Pheochromocytoma and paraganglioma (PCPG) (Figure S6F-M). In uveal melanoma (UVM), a high copy number of the three tested genes, DNAJB12, DNAJB14, and SGTA, are associated with poor prognosis and poor survival (Figure S6I, S6J, and S6M). The high copy number of DNAJB12, DNAJB14, and SGTA is also associated with poor prognosis in many other cancer types but with low significant scores. More data is needed to make significant differences (TCGA database). We suggest that the high expression of DNAJB12/14 and SGTA in those cancer types may account for the poor prognosis by inducing ERCYS and inhibiting pro-apoptotic signaling, increasing cancer cells' fitness.

      6- What do the authors mean by "just by mass action"?

      Answer: Mass action means increasing the amount of the protein (overexpression). We corrected this in the main text to overexpression.

      7- Figure 3C: Should be labelled to indicate membrane and cytosolic fraction. The AGR2 blot in the left part is not publication quality and should be replaced.

      Answer: We added the labelling to indicate cytosolic and membrane fractions to Figure 3C. We re-blotted the AGR2, new blot of AGR2 was added.

      8- What could be the reason for the fact that DNAJB12 is necessary and sufficient for ERCYS, while DNAJB14 is only necessary?

      Answer: Because of their very high homology, we speculate that the two proteins have partial redundancy. Partial because we believe that some of the roles of DNAJB12 cannot be carried by DNAJB14 in its absence. Although they are highly homologous, we expect that they probably have different affinities in recruiting other factors that are necessary for the reflux of proteins.

      We further developed around this point in the discussion and the main text.

      9- Figure 5A: Is the interaction between SGTA and JB12 UPR-independent?HCS70 seems to show only background binding. The interaction of JB12 with SGTA is not convincing. A better blot is needed.

      Answer: In the conditions of Figure 5A, we did not observe any induction of the UPR (Figure S3C-D). Thus, we concluded that in those condition of overexpression, DNAJB12 interacts with SGTA in UPR independent manner.

      We repeated this experiment another 3 times with very high number of cells (2X15cm2 culture dishes for each condition) and instead of coimmunoprecipitating with DNAJB12 antibodies we IP-ed with FLAG-beads, the results are very clear as shown in the new Figure 5A compared to Figure S5A.

      10- Figure 5B: the expression of DNAJB14 was induced by Tg50, but not by Tg25 or Tm. However, the authors have not commented on this. This should be mentioned in the text and discussed.

      Answer: In most of the experiments we did not see an increase in DNAJB14 upon ER stress except in this replicate. To be sure we looked at the DNAJB14 levels upon ER stress by protein and qPCR experiment as shown in new (in the Input of Figure 5 and Figure S5) and (Figure S5H-I). We also added new IP experiments in Figure 5 and Figure S5.

      11- Figure 6A: Why is a double knockdown important at all? DNAJB14 does not seem to do much at all (neither in overexpression nor with single knockdown).

      Answer: the data shows that DNAJB12 can compensate for the lack of DNAJB14 while DNAJB14 can only partially compensate for some of the DNAJB12 functions. DNAJB12 could have higher affinity to recruit other factor needed for the reflux process and thus the impact of DNAJB12 is higher. In summary, neither DNAJB12 or DNAJB14 is essential in the single knockdown which means that they compensate for each other. In the overexpression experiment, it is enough to have the endogenous DNAJB14 for the DNAJB12 activity. When DNAJB14 is overexpressed at very high levels, we believe that it binds to some factors that are needed for proper DNAJB12 activity (Figure 4 showing that the WT-DNAJB14 inhibits ER-stress induced ER protein reflux when overexpressed). We believe that DNAJB14 is important because only when we knock both DNAJB12 and DNAJB14 we see an effect on the ER-protein reflux. DNAJB14 is part of a complex of DNAJB12/HSC70 and DSGTA.

      (DNAJB12 is sufficient while DNAJB14 is not- please refer to point #8 above).

      **Referees cross-commenting**

      I agree with the comments raised by reviewer 1 about the manuscript. I also agree with the points written in this consultation session. In my opinion, the comments of reviewer 2 are phrased in a harsh tone and thus the reviewer reaches the conclusion that there are "serious" problems with this manuscript. However, I think that the authors could address many of the points of this reviewer in a matter of 3 months easily. For instance, it is easy to control for the expression levels of exogenous wild type and mutant D12 and compare it to the endogenous one (point 3). This is a very good point of this reviewer and I agree with this experiment. Likewise, it is easy to provide data about the levels of AGR2 to address the concern whether its synthesis is affected by D12 and D14 overexpression. Again, an excellent suggestion, but no reason for rejecting the story. As for not citing the literature, I think this can also easily be addressed and I am sure that this is just an oversight and no ill intention by the authors. __Overall, I am unable to see why the reviewer reaches such a negative verdict about this work. With proper revisions that might take 3 months, I think the points of all reviewers can be addressed. __

      Reviewer #1 (Significance (Required)):

      Significance: The strength of the work is that it provides further mechanistic insight into a novel cellular phenomenon (ERCYS). The functions for DNAJB12&14 are unprecedented and therefore of great interest for the proteostasis community. Potentially, the work is also of interest for cancer researchers, who might capitalize of the ERCYS to establish DNAJB12/14 as novel therapeutic targets. The major weaknesses are as follows: (i) the work is limited to a single cell line. To better probe the cancer relevance, the work should have used at least a panel of cell lines from one (or more) cancer entity. Ideally even data from patient derived samples would have been nice. Having said this, I also appreciate that the work is primarily in the field of cell biology and the cancer-centric work could be done by others. Certainly, the current work could inspire cancer specialists to explore the relevance of ERCYS. (ii) No physiological or pathological condition is shown where DNAJB12 is induced or depleted.

      Answer: We previously showed that ERCYS is conserved in many different cell lines including A549, MCF7, GL-261, U87, HEK293T, MRC5 and others and is also conserved in murine models of GBM (GL-261 and U87 derived tumors) and human patients with GBM (Sicari et al. 2021). Here, we tested the reflux process and the IP experiments in many different cell lines including A549, MCF-7, PC3 and Trex-293 cells. We also added new fractionation experiment in DNAJB12 and DNAJB14 -depleted MCF-7, PC3 and A549 cells. We added all those data to the revised version.

      We also added survival curves from the TCGA database showing that high copy number of DNAB12, DNAJB14 and SGTA are associated with poor prognosis compared to conditions where DNAJB12, DNAJB14, and SGTA are at low copy number (Figure S6E-M). Finally, we included immunofluorescent experiment to show that the interaction between the refluxed AGR2 and the cytosolic SGTA occurs in tumors collected from patients with colorectal cancer patients (Figure S5F-G) compared to non-cancerous tissue.

      This study is highly significant and is relevant not only to cancer but for other pathways that may behave in similar manner. For instance, DNAJB12 and DNAJB14 are part of the mechanism that is used by non-envelope viruses to escape the ER to the cytosol. Thus, the role of those DNAJB proteins seems to be mainly in the reflux of functional (not misfolded) proteins from the ER to the cytosol. We reported earlier that the UDP-Glucose-Glucosyl Transferase 1 (UGGT1) is also expelled during ER stress. UGGT1 is important because it is redeploy to the cytosol during enterovirus A71 (EA71) infection to help viral RNA synthesis (Huang et al, 2017). This redeployment of EAA71 is similar to what happens during the reflux process because on one hand, UGGT1 exit the ER by an ER stress mediated process (Sicari et al. 2021) and it is also a functional in the cytosol as a proteins which help viral RNA synthesis ((Huang et al, 2017). All those data showing that there is more of DNAJB12, DNAJB14, DNAJC14, DNAJC30 and DNAJC18 that still needs to be explored in addition to what is published. Thus, we suggest that viruses hijacked this evolutionary conserved machinery and succeeded to use it in order to escape the ER to the cytosol in a manner that depends on all the component needed for ER protein reflux.

      Reviewer #2 (Evidence, reproducibility and clarity (Required)):

      The authors present a study in which they ascribe a role for a complex containing DNAJB12/14-Hsc70-SGTA in facilitating reflux of a AGR2 from the ER to cytosol during ER-stress. This function is proposed to inhibit wt-P53 during ER-stress.

      Concerns: 1. The way the manuscript is written gives the impression that this is the first study about mammalian homologs of yeast HLJ1, while there are instead multiple published papers on mammalian orthologs of HLJ1. Section 1 and Figure 1 of the results section is redundant with a collection of previously published manuscripts and reviews. The lack of proper citation and discussion of previous literature prevents the reader from evaluating the results presented here, compared to those in the literature.

      Answer: We highly appreciate the reviewer’s comments. This paper is not to show that DNAJB12 and DNAJB14 are the orthologues of HLJ-1 but rather to show that DNAJB12 and DNAJB14 are part of a mechanism that we recently discovered and called ERCYS that cause proteins to be refluxed out of the ER. A mechanism that is regulated in by HLJ-1 in yeast. ERCYS is an adaptive and pro-survival mechanism that results in increased chemoresistance and survival in cancer cells. The papers that reviewer #2 refer to are the ones that report DNAJB12 can replace some of the ER-Associated Degradation (ERAD) functions of HLJ-1 in degradation of membranal proteins such as CFTR. These two mechanism are totally different and the role of the yeast HLJ-1 in degradation of CFTR is not needed for ERCYS. This is because we previously showed that the role of the yeast HLJ-1 and probably its orthologues in ERCYS is independent of their activity in ERAD(Igbaria et al. 2019). Surprisingly, the role of HLJ-1 in refluxing the ER proteins is not only independent of the reported ERAD-functions of HLJ1 and the mammalian DNAJBs but rather proceeds more rigorously when the ERAD is crippled (Igbaria et al. 2019). This role of DNAJBs is unique in cancer cells and is responsible in regulating the activity of p53 during the treatment of DNA damage agents.

      In our current manuscript we show by similarity, functionality, and topological orientation, that DNAJB12 and DNJB14 may be part of a well conserved mechanism to reflux proteins from the ER to the cytosol. A mechanism that is independent of DNAJB12/14’s reported activity in ERAD(Grove et al. 2011; Yamamoto et al. 2010; Youker et al. 2004). In addition, DNAJB12 and DNAJB14 facilitate the escape of non-envelope viruses from the ER to the cytosol in similar way to the reflux process(Goodwin et al. 2011; Igbaria et al. 2019; Sicari et al. 2021). All those data show that HLJ-1 reported function may be only the beginning of our understanding on the role that those orthologues carry and that are different from what is known about their ERAD function.

      Action: We added the references to the main text and discussed the differences between the reported DNAJB12 and HLJ-1 functions to the function of DNAJB12, DNAJB14 and the other DNAJ proteins in the reflux process. We also developed around this in the discussion.

      The conditions used to study DNAJB12 and DNAJ14 function in AGR2 reflux from the ER do not appear to be of physiological relevance. As seen below they involve two transfections and treatment with two cytotoxic drugs over a period of 42 hours. The assay for ERCY is accumulation of lumenal ER proteins in a cytosolic fraction. Yet, there is no data or controls that describe the path taken by AGR2 from the ER to cytosol. It seems like pleotropic damage to the ER due the experimental conditions and accompanying cell death could account for the reported results?

      Transfection of cells with siRNA for DNAJB12 or DNAJB14 with a subsequent 24-hour growth period.

      Transfection of cells with a p53-lucifease reporter.

      Treatment of cells with etoposide for 2-hours to inhibit DNA synthesis and induce p53. D. Treatment of cells for 16 hours with tunicamycin to inhibit addition of N-linked glycans to secretory proteins and cause ER-stress.

      Subcellular fractionation to determine the localization of AGR2, DNAJB11, and HYOU1

      KD of DNAJB12 or DNAJB14 have modest if any impact on AGR2 accumulation in the cytosol. There is an effect of the double KD of DNAJB12 or DNAJB14 on AGR2 accumulation in the cytosol. Yet there are no western blots showing AGR2 levels in the different cells, so it is possible that AGR2 is not synthesized in cells lacking DNAJB12 and DNAKB14. The lack of controls showing the impact of single and double KD or DNAJB12 and DNAJB14 on cell viability and ER-homeostasis make it difficult to interpret the result presented. How many control versus siRNA KD cells survive the protocol used in these assays?


      Answer: Despite the long protocol we see differences between the control cells and the DNAJB-silenced cells in terms of the quantity of the refluxed proteins to the cytosol. The luciferase construct was used to assess the activity of p53 so the step of the second transfection was used only in experiments were we assayed the p53-luciferase activity. The rest of the experiments especially those where we tested the levels of p53 and P21 levels, were performed with one transfection. Moreover, all the experiments with the subcellular protein fractionation were performed after one transfection without the second transfection of the p53-Luciferase reporter. Finally, the protocol of the subcellular protein fractionation requires first to trypsinize the cells to lift them up from the plates, at the time of the experiment the cells were almost at 70-80% confluency and in the right morphology under the microscope.

      Here, we performed XTT assay and Caspase-3 assay to asses cell death at the end of the experiment and before the fractionation assay. We did not observe any differences at this stage between the different cell lines (Figure-RV1 for reviewers Only). This can be explained by the fact that we use low concentrations of Tm and Tg for short time of 16 hour after the pulse of etoposide.

      Finally, the claim that and ER-membrane damage result in a mix between the ER and cytosolic components is not true for the following reasons: (1) In case of mixing we would expect that GAPDH levels in the membrane fraction will be increased and that we do not see, and (2) we used our previously described transmembrane-eroGFP (TM-eroGFP) that harbors a transmembrane domain and is attached to the ER membrane facing the ER lumen. The TM-eroGFP was found to be oxidized in all conditions tested. Those data argue against a rupture of the ER membrane which can results in a mix of the highly reducing cytosolic environment with the highly oxidizing ER environment by the passage of the tripeptide GSH from the cytosol to the ER. All those data argue against (1) cell death, and (2) rupture of the ER membrane. Figure RV1 Reviewers Only.

      Moreover, as it is shown in Figure S2, AGR2 is found in the membrane fraction in all the four different knock downs, thus it is synthesized in all of them. Moreover, we assayed the mRNA levels of AGR2 in all the knockdowns and we so that they are at the same levels in all the 4 different conditions and still AGR2 mRNA levels increase upon ER stress in all of the 4 knockdown cells in different backgrounds (Figure S2F-N).

      In Figure 3 the authors overexpress WT-D12 and H139Q-D12 and examine induction of the p53-reporter. There are no western blots showing the expression levels of WT-D12 and H139Q-D12 relative to endogenous DNAJB12. HLJ1 stands for high-copy lethal DnaJ1 as overexpression of HLJ1 kills yeast. The authors present no controls showing that WT-D12 and H139-D12 are not expressed at toxic levels, so the data presented is difficult to evaluate.

      Answer: The expression levels of the overexpression of DNAJB12 and DNAJB14 were present in the initial submission of the manuscript as Figure S3A and S3B. The data showing the relationship between the expression degree and the viability were also included in the initial submission as Figure S3C (Now S3H).

      There is no mechanistic data used to help explain the putative role DNAJB12 and DNAJB14 in ERCY? In Figure 4, why does H139Q JB12 prevent accumulation of AGR2 in the cytosol? There are no westerns showing the level to which DNAJB12 and DNAJB14 are overexpressed.


      Answer: The data showing the levels of DNAJB12 compared to the endogenous were present in the initial submission as Figure S3A and S3B.

      We suggest a mechanism by which DNAJB12 and DNAJB14 interact (Figure 5 and Figure S5) and oligomerize to expel those proteins in similar way to expelling non-envelope viruses to the cytosol. Thus, when expressing the mutant DNAJB12 H139Q may indicate that the J-domain dead-mutant can still be part of the complex but affects the J-domain activity in this oligomer and thus inhibit ER-protein reflux. In other words, we showed that the H139Q exhibits a dominant negative effect when overexpressed. Moreover, here we added another IP experiment in the D12/D14-DKD cells to show that in the absence of DNAJB12 and DNAJB14, SGTA cannot bind the ER-lumenal proteins because they are not refluxed (Figure 5 and Figure S5). Those data indicate that in order for SGTA bind the refluxed proteins they have to go through the DNAJB12 and DNAJB14 and their absence this interaction does not occur. This explanation was also present in the discussion of the initial submission.

      Mechanistically, we show that AGR2 interacts with DNAJB12/14 which are necessary for its reflux. This mechanism involves the functionality of cytosolic HSP70 chaperones and their cochaperones (SGTA) proteins that are recruited by DNAJB12 and 14. This mechanism is conserved from yeast to mammals. Moreover, by using the alpha-fold prediction tools, we found that AGR2 is predicted to interact with SGTA in the cytosol by the interaction between the cysteines of SGTA and AGR2 in a redox-dependent manner.

      **Referees cross-commenting**

      __ __ I appreciate the comments of the other reviewers. I agree that the authors could revise the manuscript. Yet, based on my concerns about the physiological significance of the process under study and lack of scholarship in the original draft, I would not agree to review a revised version of the paper.

      Answer: Regards the physiological relevance, we showed in our previous study (Sicari et al. 2021) how relevant is ERCYS in human patients of GBM and murine model of GBM. ERCYS is conserved from yeast to human and is constitutively active in GL-261 GBM model, U87 GBM model and human patients with GBM (Sicari et al. 2021). Here, extended that to other tumors and showed that DNAJB12, DNAJB14 and SGTA high levels are associated with poor prognosis in many cancer types (Figure S6). We also show some data from to show the relevance and added data showing the interaction of SGTA with AGR2 in CRC samples obtained from human patients compared to healthy tissue (Figure S5). This study is highly significant and is relevant not only to cancer but for other pathways that may behave in similar manner. For instance, DNAJB12 and DNAJB14 are part of the mechanism that is used by non-envelope viruses to escape the ER to the cytosol. Thus, the role of those DNAJB proteins seems to be mainly in the reflux of functional (not misfolded) proteins from the ER to the cytosol. We reported earlier that the UDP-Glucose-Glucosyl Transferase 1 (UGGT1) is also expelled during ER stress. UGGT1 is important because it is redeploy to the cytosol during enterovirus A71 (EA71) infection to help viral RNA synthesis (Huang et al, 2017). This redeployment of EAA71 is similar to what happens during the reflux process because on one hand, UGGT1 exit the ER by an ER stress mediated process (Sicari et al. 2021) and it is also a functional in the cytosol as a proteins which help viral RNA synthesis ((Huang et al, 2017). All those data showing that there is more of DNAJB12, DNAJB14, DNAJC14, DNAJC30 and DNAJC18 that still needs to be explored in addition to what is published. We suggest that viruses hijacked this evolutionary conserved machinery and succeeded to use it in order to escape.

      We appreciate the time spent to review our paper and we are sorry that the reviewer reached such verdict that is also not understood by the other reviewers. Most of the points raised by reviewer 2 were already addressed and explained in the initial submission, anyways we appreciate the time and the comments of reviewer #2 on our manuscript.

      Reviewer #2 (Significance (Required)):

      Overall, there are serious concerns about the writing of this paper as it gives the impression that it is the first study on higher eukaryotic and mammalian homologs of yeast HLJ1. The reader is not given the ability to compare the presented data to related published work. There are also serious concerns about the quality of the data presented and the physiological significance of the process under study. In its present form, this work does not appear suitable for publication.

      Answer: Again we thank reviewer #2 for giving us the opportunity to explain how significant is this manuscript especially for people who are less expert in this field. The significance of this paper (1) showing a the unique role of DNAJB12 and DNAJB14 in the molecular mechanism of the reflux process in mammalian cells (not their role in ERAD), (2) showing the implication of other cytosolic chaperones in the process including HSC70 and SGTA (3), our alpha-fold prediction show that this process may be redox dependent that implicate the cysteines of SGTA in extracting the ER proteins, (4) overexpression of the WT DNAJB12 is sufficient to drive this process, (5) mutation in the HPD motif prevent the reflux process probably by preventing the binding to the cytosolic chaperones, and (6) we need both DNAJB12 and DNAJB14 in order to make the interaction between the refluxed ER-proteins and the cytosolic chaperones occur.

      In Summary, this study is highly significant in terms of physiology, we previously reported that ERCYS is conserved in mammalian cells and is constitutively active in human and murine tumors (Sicari et al. 2021). Moreover, DNAJB12 and DNAJB14 are part of the mechanism that is used by non-envelope viruses to escape the ER to the cytosol in a mechanism that is similar to reflux process (Goodwin et al. 2011; Goodwin et al. 2014). Thus, the role of those DNAJB proteins seems to be mainly in the reflux of functional proteins from the ER to the cytosol, viruses used this evolutionary conserved machinery and succeeded to use in order to escape. This paper does not deal with the functional orthologues of the HLJ-1 in ERAD but rather suggesting a mechanism by which soluble proteins exit the ER to the cytosol.

      __Reviewer #3 (Evidence, reproducibility and clarity (Required)):____ __

      Summary: Reflux of ER based proteins to the cytosol during ER stress inhibits wt-p53. This is a pro-survival mechanism during ER stress, but as ER stress is high in many cancers, it also promotes survival of cancer cells. Using A549 cells, Dabsan et al. demonstrate that this mechanism is conserved from yeast to mammalian cells, and identify DNAJB12 and DNAJB14 as putative mammalian orthologues of yeast HLJ1.

      This paper shows that DNAJB12 and 14 are likely orthologues of HLJ1 based on their sequences, and their behaviour. The paper develops the pathway of ER-stress > protein reflux > cytosolic interactions > inhibition of p53. The authors demonstrate this nicely using knock downs of DNAJB12 and/or 14 that partially blocks protein reflux and p53 inhibition. Overexpression of WT DNAJB12, but not the J-domain inactive mutant, blocks etoposide-induced p53 activation (this is not replicated with DNAJB14) and ER-resident protein reflux. The authors then show that DNAJB12/14 interact with refluxed ER-resident proteins and cytosolic SGTA, which importantly, they show interacts with the ER-resident proteins AGR2, PRDX4 and DNAJB11. Finally, the authors show that inducing ER stress in cancer cell lines can increase proliferation (lost by etoposide treatment), and that this is partially dependent on DNAJB12/14.

      This is a very interesting paper that describes a nice mechanism linking ER-stress to inhibition of p53 and thus survival in the face of ER-stress, which is a double edged sword regarding normal v cancerous cells. The data is normally good, but the conclusions drawn oversimplify the data that can be quite complex. The paper opens a lot of questions that the authors may want to develop in more detail (non-experimentally) to work on these areas in the future, or alternatively to develop experimentally and develop the observations further. There are only a few experimental comments that I make that I think should be done to publish this paper, to increase robustness of the work already here, the rest are optional for developing the paper further.

      We thank the reviewer for his/her positive comments His/her comments contributed to make our manuscript stronger.

      __Major comments:____ __

      1. Number of experimental repeats must be mentioned in the figure legends. Figures and annotations need to be aligned properly

      __Answer____: __All experiments were repeated at least 3 times. We added the number of repeats on each figure in the figures legends

      Results section 2:

      No intro to the proteins you've looked at for relocalization. Would be useful to have some info on why you chose AGR2. Apart from them being ER-localized, do they all share another common characteristic? Does ability to inhibit p53 vary in potency?

      Answer: We previously showed that AGR2 is refluxed from the ER to the cytosol to bind and inhibit wt-p53 (Sicari et al. 2021). Here, we used AGR2 because, (1) we know that AGR2 is refluxed from the ER to the cytosol, and (2) we know which novel functions it gains in the cytosol so we are able to measure and provide a physiological significance of those novel functions when the levels of DNAJB12 and DNAJB14 are altered. Moreover, we used DNAJB11 (41 kDa) and HYOU1 (150 kDa) proteins to show that alteration in DNAJB12 or DNAJB14 prevent the reflux small, medium and large sized proteins. We added a sentence in the discussion stating that DNAJB12/14 are responsible for the reflux of ER-resident proteins independently of their size. We also added in the result section that we are looking at proteins of different sizes and activities.


      What are the roles of DNAJB12/14 if overexpression can induce reflux? Does it allow increased binding of an already cytosolic protein, causing an overall increase in an interaction that then causes inhibition of p53? What are your suggested mechanisms?

      Answer: Previously it was reported that over-expression of DNAJB12 and DNAJB14 tend to form membranous structures within cell nuclei, which was designate as DJANGOS for DNAJ-associated nuclear globular structures(Goodwin et al. 2014). Because those structures which contain both DNAJB12 and DNAJB14 also form on the ER membrane (Goodwin et al. 2014), we speculate that during stress DNAJB12/14 overexpression may facilitate ERCYS. Interestingly, those structures contain Hsc70 and markers of the ER lumen, the nuclear and ER and nuclear membranes (Goodwin et al. 2014).

      The discussion was edited accordingly to further strengthen and clarify this point

      Fig3: A+B show overexpression of individual DNAJs but not combined. As you go on to discuss the effect of the combination on AGR2 reflux, it would be useful to include this experimentally here.

      Answer: This is a great idea, we tried to do it for long time. Unfortunately when we used cells overexpress DNAJB12 under the doxycycline promoter and transfect with DNAJB14 plasmid expressing DNAJB14 under the CMV promoter, most of the cells float within 24 hours compared to cells transfected with the empty vector alone or with DNAJB14-H136Q. We also did overexpression of DNAJB14 in cells with DNAJB12 conditional expression and also were lethal in Trex293T cells and A549-cells.

      Fig 3C: Subfractionation of cells shows AGR2 in the cytosol of A549 cells. The quality of the data is good but the bands are very high on the blot. For publication is it possible to show this band more centralized so that we are sure that we are not missing bands cut off in the empty and H139Q lanes?

      Also, you have some nice immunofluorescence in the 2021 EMBO reports paper, is it possible to show this by IF too? It is not essential for the story, but it would enrich the figure and support the biochemistry nicely. Also it is notable that the membrane fraction of the refluxed proteins doesn't appear to have a decrease in parallel (especially for AGR2). Is this because the % of the refluxed protein is very small? Is there a transcriptional increase of any of them (the treatments are 12+24 h so it would be enough time)? This could be a nice opportunity to discuss the amount of protein that is refluxed, whether this response is a huge emptying of the ER or more like a gentle release, and also the potency of the gain of function and effect on p53 vs the amount of protein refluxed. This latter part isn't essential but it would be a nice element to expand upon.

      Answer: We re-blotted the AGR2 again, new blot of AGR2 was added. More blots also are added in Figure S2, the text is edited accordingly.

      In new Figure S5 we added immunofluorescence experiment from tumors and non-tumors tissues obtained from Colorectal cancer (CRC) patients showing that the interaction between SGTA and the refluxed AGR2 also occurs in more physiological settings. It is also to emphasize that the suggested mechanism that implicates SGTA is also valid in CRC tumors.

      We performed qPCR experiments in control, DNAJB12-KD, DNAJB14-KD and in the DNAJB12/DNAJB14 double knock down cells (in both A549 and PC3 cells) to follow the mRNA levels of DNAJB11. As shown in the Figure S2F-N, there is no increase in the mRNA levels of DNAJB11, AGR2 or HYOU1 in the different cells in normal (unstressed conditions). Upon ER stress with tunicamycin or thapsigargin there is a little increase in the mRNA levels of HYOU1 and AGR2 but in DNAJB11 mRNA levels. On the other hand, we also performed western blot analysis and we did not detect any difference between the different knockdown cells when we analyzed the levels of DNAJB11 compared to GAPDH. Those data are now added to Figure S2F-N. We must note that in AGR2 and HYOU1 are induced at the mRNA as a result of ER stress. The data with the overexpression of DNAJB12 and DNAJB14 are important control experiment where we show a reflux when DNAJB12 is overexpressed without inducing the ER stress (Figure 3, Figure 4, and Figure S3). In those conditions no induction of AGR2, HYOU1 or DNAJB11 were observed. Those results argue against the reflux as a result of protein induction and the increase in the proteins levels.

      The overall protein levels in steady state are function of how much proteins are made, degraded and probably secreted outside the cell. We do see in Figure S2 under ER stress there are some differences in the levels of the mRNA, moreover, from our work in yeast we showed that the expelled proteins have very long half-life in the cytosol (Igbaria et al. 2019). Because it is difficult to assay how many of the mRNA is translated and how much of it is stable/degraded and the stability of the cytosolic fraction vs the ER, it is hard to interpret on the stability and the levels of the proteins.

      Those data are now added to the manuscript, the text is edited accordingly.

      You still mention DNAJB12 and 14 as orthologues, even though DNAJB14 has no effect on p53 activity when overexpressed. Do you think that this piece of data diminishes this statement?

      Answer: The fact that DNAJB12 and DNAJB14 are highly homologous and that only the double knockdown has a great effect on the reflux process may indicate that they are redundant. Moreover, because only DNAJB12 is sufficient may indicate that some of DNAJB12 function cannot be carried by DNAJB14. In one hand they share common activities as shown in the double knock down and on the other hand DNAJB12 has a unique function that may not be compensated by DNAJB14 when overexpressed.

      __ __ Fig 3D/F: Overexpression of DNAJB14 induces reflux of DNAJB11 at 24h, what does this suggest? Does this indicate having the same role as DNAJB12 but less potently? What's your hypothesis?

      Answer: ERCYS is new and interesting phenomenon and the redistribution of proteins to the cytosol has been documented lately by many groups. Despite that we still do not know what is the specificity of DNAJB12 and DNAJB14 to the refluxed proteins. DNAJB11 is glycosylated protein and now we are testing whether other glycosylated proteins prefer the DNAJB14 pathway or not. This data is beyond the scope of this paper

      "This suggests that the two proteins may have different functions when overexpressed, despite their overlapping and redundant functions" What does it suggest about their dependence on each other? If overexpression of WT DNAJB12 inhibits Tg induced reflux, is it also blocking the ability of DNAJB14 to permit flux?

      Answer: We hypothesize that it is all about the stichometry and the ratios between proteins. When we overexpress DNAJB14 (the one that is not sufficient to cause reflux it may hijack common components and factor by non-specifically binding to them. Those factors may be needed for DNAJB12 to function properly (Like the dominant negative effect of the DNAJB12-HPD mutant for instance). On the other hand, DNAJB12 may have higher affinity for some cytosolic partner and thus can do the job when overexpressed. Here, we deal with the DNAJB12/DNAJB14 as essential components of the reflux process, yet we need to identify the interactome of each of the proteins during stress and the role of the other DNAJ proteins that also share some of the topological and structural similarity to DNAJB12, DNAJB14 and HLJ-1 (DNAJC30, DNAJC14, and DNAJC18). We edited the text accordingly and integrated this in the discussion.

      __ __ Fig 4: PDI shown in blots but not commented on in text. Then included in the schematics. Please comment in the text.

      Answer: We commented PDI in the text.

      Fig 4F: Although the quantifications of the blots look fine, the blot shown does not convincingly demonstrate this data for AGR2. The other proteins look fine, but again it could be useful to see the individual means for each experiment, or the full gels for all replicates in a supplementary figure.

      Answer: the other two repeats are in Figure S4

      __ __Results section 3

      Fig 5A, As there is obviously a difference between DNAJB12/14 it would be useful to do the pulldown with DNAJB14 too. Re. HSC70 binding to DNAJB12 and 14, the abstract states that DNAJB12/14 bind HSC70 and SGTA through their cytosolic J domains. Fig 5 shows pulldowns of DNAJB12 with an increased binding of SGTA in FLAG-DNAJB12 induced conditions, but the HSC70 band does not seem to be enriched in any of the conditions, including after DNAJB12 induction. This doesn't support the statement that DNAJB12 binds HSC70. In fact, in the absence of a good negative control, this would suggest that the HSC70 band seen is not specific. There is also no data to show that DNAJB14 binds HSC70. I recommend including a negative condition (ie beads only) and the data for DNAJB14 pulldown.

      Answer: In Figure 5A we used the Flp-In T-REx-293 cells as it is easier to control and to tune up and down the expression levels of DNAJB12 and DNAJB14. According to new Figure S5A, DNAJB12 binds at the basal levels to HSC70 all the time. It was also surprising for us not to see the differences in the overexpression and we relate that to the fact that all the HSC70 are saturated with DNAJB12. In order to better assay that we repeated the IP in Figure 5A but instead of the IP with DNAJB12, we IP-ed with FLAG antibodies to selectively IP the transfected DNAJB12. As shown in the new Fig 5A, the increase of DNAJB12-FLAG is accompanied with an increase in the binding of HSC70.

      We further tested the interaction between DNAJB12, DNAJB14 and HSC70 during ER stress in cancer cells. In those cells we found that DNAJB12 and DNAJB14 bind to HSC70 and they recruit SGTA upon stress. We also tested the binding between DNAJB12 and DNAJB14, in unstressed conditions, there was a basal binding between both, this interaction was stronger during ER stress. Those data are now added to Figure 5 and Figure S5 and the discussion was edited accordingly.

      The binding of DNAJB12 to SGTA under stress conditions in Fig5B looks much more convincing than SGTA to DNAJB12 in Fig 5A. Bands in all blots need to be quantified from 3 independent experiments, and repeated if not already n=3. If this is solely a technical difference, please explain in the text.

      The conclusions drawn from this interaction data are important and shold be elaborated upon to support th claims made in the paper. The authors may also chose to expand the pulldowns to demonstrate their claims made on olidomerisation of DNAJB12 and 14 here. It is also clear that the interaction data of the SGTA with ER-resident proteins AGR2, PRDX4 and DNAJB11 is strong. The authors may want to draw on this in their hypotheses of the mechanism. I would imagine a complex such as DNAJB14/DNAJB12 - SGTA - AGR2/PRDX4/DNAJB11 would be logical. Have any experiments been performed to prove if complexes like this would form?

      Answer: In Figure 5A we used the Flp-In T-REx-293 cells as it is easier to control and to tune up and down the expression levels of DNAJB12 and DNAJB14. T-REx-293 are highly sensitive to ER stress, they do not die (as we did not observe apoptosis markers to be elevated) but they float and can regrow after the stress is gone. In Figure 5B we are using ER stress without the need to express DNAJB12 in A549 cell line. In order to further verify those data, we repeated the IP in another cell line as well to confirm the data in 5B. We also repeated the IP in 5A with anti-FLAG antibody to improve the IP and to specifically map he interaction with the overexpressed FLAG-DNAJB12 (discussed above). All experiments were done in triplicates and added to Figure 5 and Figure S5.

      We agree with the reviewer on the complex between the refluxed proteins and SGTA. We believed that SGTA may form a complex with other refluxed ER-proteins but we were unable to see an interaction between AGR2-DNAJB11 in the cytosolic fraction or between AGR2-PRDX4 in the conditions tested in the cytosolic fraction. We could not do this in the whole cell lysate because those proteins bind each other in the ER. Finally, our structural prediction using Alpha-fold suggests that the interaction between SGTA and the refluxed AGR2 (and probably others) is redox depending and that it requires disulfide bridge between cysteine 81 on AGR2 and cysteine 153 on SGTA. Thus, we hypothesize that SGTA binds one refluxed protein at the time.

      We repeated the figure with improvement: (1) using more cells in order to increase the amount of IP-ed proteins and to overcome the problem of the faint bands, (2) performing the IP with the FLAG antibodies instead of the DNAJB12 endogenous antibodies.

      Fig 5B: It is clear that DNAJB12 interacts with SGTA. The authors state that DNAJB14 also interacts with SGTA under normal and stress conditions, but the band in 25/50 Tg is very feint. Why would there be stronger binding at the 2 extremes than during low stress induction? In the input, there is a much higher expression of DNAJB14 in 50 Tg. What does this say about the interaction? Is there an effect of ER stress on DNAJB14 expression? A negative control should be included to show any background binding, such as a "beads only" control

      __Answer: __DNAJB14 does not change with ER stress as shown in the Ips (Input) and in the qPCR experiment in Figure S5I. We added beads only control, we also added new Ips to assess the binding between DNAJB14 and DNAJB12, and between DNAJB14-SGTA. All the new Ips and controls now added as Figure 5 and Figure S5.

      Fig 5C data is sound, although a negative control should be included.

      Answer: Negative control was added in Figure S5.

      __Results section 4____ __

      Fig 6A-B: Given that there is the complexity of overexpression v KD of DNAJB12 v 14 causing similar effects on p53 actvity (Fig 2 v 3), it would be interesting to see whether the effect of overexpression mirrors the results in Fig 6A. Is it known what SGTA overexpression does (optional)?

      Answer: In the overexpression system, cells overexpressing DNAJB12 start to die between 24-48 hours as shown in Figure S3C. Thus, it is difficult to assay the proliferation of these cells in those conditions. On the other hand, overexpression of Myc-tagged SGTA in A549 cells, MCF7 or T-ReX293 did not show any reflux of ER-proteins to the cytosol and it didn’t show any significant changes in the proliferation index (Figure Reviewers only RV2).

      Fig 6D: resolution very low

      Answer: Figure 6D was changed

      __ __ Fig 6C-D: There is an interesting difference though between the proposed cytosolic actions of the refluxed proteins. You show that AGR2, PRDX4 and DNAJB11 all bind to SGTA in stress conditions, but in the schematics you show: DNAJB11 binding to HSC70 through SGTA (not shown in the paper), then also PDIA1, PDIA3 binding to SGTA and AGR2 binding to SGTA. What role does SGTA have in these varied reactions? Sometimes it is depicted as an intermediate, sometimes a lone binder, what is its role as a binder? It should be clarified which interactions are demonstrated in the paper (or before) and which are hypothesized in a graphical way (eg. for hypotheses dotted outlines or no solid fill etc). The schematics also suggest that DNAJB14 binding to HSC70 and SGTA is inducible in stress conditions, as is PDIA3, which is not shown in the paper. Discussion "In cancer cells, DNAJB12 and DNAJB14 oligomerize and recruit cytosolic chaperones and cochaperones (HSC70 and SGTA) to reflux AGR2 and other ER-resident proteins and to inhibit wt-p53 and probably different proapoptotic signaling pathways (Figure 5, and Figure 6C-6D)." You havent shown oligomerisation between DNAJB12/14. Modify the text to make it clear that it is a hypothesis.

      Answer: We removed “oligomerize” from the text and added that it as a hypothesis. Figure (C-D) also were changed to be compatible with the text.

      Minor comments:

      __ __ It would be useful to have page or line numbers to help with document navigation, please include them. Typos and inconsistency in how some proteins are named throughout the manuscript

      Answer: Page numbers and line numbers are added. Typos are corrected

      Title: Include reference to reflux. Suggest: "chaperone complexes (?proteins) reflux from the ER to cytosol..." I presume it would be more likely that the proteins go separately rather than in complex. Do you have any ideas on the size range of proteins that can undergo this process?

      Answer: this is true, proteins may cross the ER membrane separately and then be in a complex with cytosolic chaperones. The title is changed accordingly. As discussed earlier, the protein we chose were of different sizes to show that they are refluxed independently of their size. Moreover, our previous work showed that the proteins that were refluxed are of different sizes. Most importantly UGGT1 (around 180 Kda) which is reported to deploy to the cytosol upon viral infection (Huang et al. 2017; Sicari et al. 2020). In this study we used AGR2 (around 19 Kda) and HYOU1 (150Kda).

      ERCY in abstract, ERCYS in intro. There are typos throughout, could be a formatting problem, please check

      Answer: Checked and corrected

      What about the selection of refluxed proteins? Is this only a certain category of proteins? Could it be anything? Have you looked at other cargo / ER resident proteins?

      __ ____Answer: __in our previous study by (Sicari, Pineau et al. 2020) we looked at many other proteins especially glycoproteins from the ER. In (Sicari, Pineau et al. 2020) we used mass spectrometry in order to identify new refluxed proteins and we found 26 new glycoprotein that are refluxed from cells treated with ER stressor and from human tissues obtained from GBM patients (Sicari, Pineau et al. 2020).

      We previously showed that AGR2 is refluxed from the ER to the cytosol to bind and inhibit p53 (Sicari, Pineau et al. 2020). Here, we selected AGR2 because we know that (1) it is refluxed, and (2) we know which novel functions it acquires in the cytosol so we are able to measure and provide a physiological significance of those novel functions when the levels of DNAJB12 and DNAJB14 are altered. Moreover, we selected DNAJB11 (41 kDa) and HYOU1 (150 kDa) proteins to show that alteration in DNAJB12 or DNAJB14 prevent the reflux small, medium and large protein (independently of their size). We also showed earlier by mass spectrometry analysis that the refluxed proteins range from small to very large proteins such as UGGT1, thus we believe that soluble ER-proteins can be substrates of ERCYS independently of their size. In the discussion, we added a note that the reflux by the cytosolic and ER chaperones operates on different proteins independently of their size.

      "Their role in ERCYS and cells' fate determination depends..." Suggest change to "Their role in ERCYS and determination of cell fate..."

      Answer: changed and corrected

      I think that the final sentence of the intro could be made stronger and more concise. There's a repeat of ER and cytosol. Instead could you comment on the reflux permitting new interactions between proteins otherwise spatially separated, then the effect on wt-p53 etc.

      Answer: The sentence was rephrased as suggested to “ In this study, we found that HLJ1 is conserved through evolution and that mammalian cells have five putative functionality orthologs of the yeast HLJ1. Those five DNAJ- proteins (DNAJB12, DNAJB14, DNAJC14, DNAJC18, and DNAJC30) reside within the ER membrane with a J-domain facing the cytosol (Piette et al. 2021; Malinverni et al. 2023). Among those, we found that DNAJB12 and DNAJB14, which are strongly related to the yeast HLJ1 (Grove et al. 2011; Yamamoto et al. 2010), are essential and sufficient for determining cells' fate during ER stress by regulating ERCYS. Their role in ERCYS and determining cells' fate depends on their HPD motif in the J-domain. Downregulation of DNAJB12 and DNAJB14 increases cell toxicity and wt-p53 activity during etoposide treatment. Mechanistically, DNAJB12 and DNAJB14 interact and recruit cytosolic chaperones (HSC70/SGTA) to promote ERCYS. This later interaction is conserved in human tumors including colorectal cancer.

      In summary, we propose a novel mechanism by which ER-soluble proteins are refluxed from the ER to the cytosol, permitting new inhibitory interactions between spatially separated proteins. This mechanism depends on cytosolic and ER chaperones and cochaperones, namely DNAJB12, DNAJB14, SGTA, and HSC70. As a result, the refluxed proteins gain new functions to inhibit the activity of wt-p53 in cancer cells. “

      __Figure legends: __

      In some cases the authors state the number of replicates, but this should be stated for all experiments. If experiments don't already include 3 independent repeats, this should be done. Check text for typos, correct letter capitalisation, spaces and random bold text (some of this could be from incompatability when saving as PDF)

      Answer: all experiments were repeated at least three times. The number of repeats is now indicated in the figure legends of each experiment. Typos and capitalization is corrected as well.

      Fig2E: scrambled not scramble siRNA

      Answer: corrected

      Fig 3: "to expel" is a term not used in the rest of the paper for reflux. Useful to remain consistent with terminology where possible

      Answer: Rephrased and corrected

      Results section 1:

      "Protein alignment of the yeast HLJ1p showed high amino acids similarity to the mammalian..."

      Answer: Rephrased to “ Comparing the amino acid sequences revealed significant similarity between the yeast protein HLJ1p and the mammalian proteins DNAJB12 and DNAJB14”

      __ __ Fig 1C: state in legend which organism this is from (presumably human)

      Answer: in Figure 1C legends it is stated that: “ the HPD motif within the J-domain is conserved in HLJ-1 and its putative human orthologs DNAJB12, DNAJB14, DNAJC14, DNAJC18, and DNAJC30.”

      Results Section 2

      "Test the two strongest hits DNAJB12/14" Add reference to previous paper showing this

      Answer: the references were added.

      __ __ "In the WT and J-protein-silenced A549 cells, there were no differences in the cytosolic enrichment of the three ER resident proteins AGR2, DNAJB11, and HYOU1 in normal and unstressed conditions (Figure 2A-C and Figure S2C)." I think that this is an oversimplification, and in your following discussion, you show this it's more subtle than this.

      Answer: We expanded on this both in the discussion and the results section.

      __ __ The text here isn't so clear: normal and unstressed conditions? Do you mean stressed? Please be careful in your phrases: "DNAJB12-silenced cells are slightly affected in AGR2 and DNAJB11 cytosolic accumulation but not HYOU1." This is the wrong way around. DNAJB12 silencing effects AGR2, not that AGR2 effects the cells (which is how you have written it). This also occurs agan in the next para:

      Answer: Normal cells are non-cancer cells. Unstressed conditions= without ER stress. The sentence was rephrased to: In the absence of ER stress, the cytosolic levels of the three ER-resident proteins (AGR2, DNAJB11, and HYOU1) were similar in wild-type and J-protein-silenced A549 cells.

      "During stress, DNAJB12/DNAJB14 double knockdown was highly affected in the cytosolic..." I think you mean it highly affected the cytosolic accumulation, not that it was affected by the cytosolic accumulation. Please change in the text

      Answer: the sentence is now rephrased to” During stress, double knockdown of DNAJB12 and DNAJB14 highly affected the cytosolic accumulation of all three tested proteins”

      __ __ "DNAJB12 and DNAJB14 are strong hits of the yeast HLJ1" Not clear, I presume you mean they are likely orthologues? Top candidates for being closest orthologues?

      Answer: this is correct, the sentence is rephrased and corrected

      __ __ Fig 2D: typos in WB labelling? I think Tm should be - - +, not - + +as it is now (if it's not a typo, you need more controls, eto alone.

      Answer: the labeling is now corrected

      Fig 2D-E-F typos for DKD? D12/D12 or D12/14?

      Answer: This is correct, thank you for pointing this out. The labeling in corrected

      __ __ "We assayed the phosphorylation state of wt- p53 and p21 protein expression levels (a downstream target of p53 signaling) during etoposide treatment." What are the results of this? Explain what Fig 2D-E shows, then build on this with the +Tm results. Results should be explained didactically to be clear.

      Answer: The paragraph was edited and we explained the results: In these conditions, we saw an increase in the phosphorylation of wt-p53 in the control cells and in cells knocked-down with DNAJB12, DNAJB14 or both. This phosphorylation increased the protein levels of p21 as well (Figure 2D-G). Tm addition to cells treated with etoposide resulted in a reduction in wt-p53 phosphorylation, and as a consequence, the p21 protein levels were also decreased (Figure 2D-G and Figure S2O). Cells lacking DNAJB12 or DNAJB14 have partial protection in wt-p53 phosphorylation and p21 protein levels. Silencing both proteins in A549 and MCF7 cells rescued wt-p53 phosphorylation and p21 levels (Figure 2D-G and Figure S2D). Moreover, similar results were obtained when we assayed the transcriptional activity of wt-p53 in cells transfected with a luciferase reporter under the p53-DNA binding site (Figure 2H). These data confirm that DNAJB12 and DNAJB14 are involved in ER protein reflux and the inhibition of wt-p53 activity during ER stress.


      "(Figure 2D- E). Cells lacking DNAJB12 and or DNAJB14 have partial protection in wt-p53 phosphorylation and p21 protein levels."

      Answer: This sentence is now removed

      You comment on p53 phosphorylation, but you haven't quantified this. This should be done, normalized to p53 levels, if you want to draw these conclusions, especially as total p53 varies between condition. Does Eto increase p53 txn? Does Tm alone increase p53 activity/phospho-p53? These are shown in the Sicari EMBO reports paper in 2021, you should briefly reference those.

      Answer: The blots are now quantified and new blot is added to Figure S2D. The Paragraph was edited and referenced to our previous paper (Sicari et al. 2021). “We then wanted to examine whether the gain of function of AGR2 and the inhibition of wt-p53 depends on the activity of DNAJB12 and DNJAB14. We assayed the phosphorylation state of wt-p53 and p21 protein expression levels (a downstream target of wt-p53 signaling) during etoposide treatment. In these conditions, there was an increase in the phosphorylation of wt-p53 in the control cells and in cells knocked down with DNAJB12, DNAJB14, or both. This phosphorylation also increases protein levels of p21 (Figure 2D-G and Figure S2O). Tm addition to cells treated with etoposide resulted in a reduction in wt-p53 phosphorylation, and as a consequence, the p21 protein levels were also decreased (Figure 2D-G and Figure S2O). Silencing DNAJB12 and DNAJB14 in A549 and MCF-7 cells rescued wt-p53 phosphorylation and p21 levels (Figure 2D-G and Figure S2O). Moreover, similar results were obtained when we assayed the transcriptional activity of wt-p53 in cells transfected with a luciferase reporter under the p53-DNA binding site (Figure 2H). In the latter experiment, etoposide treatment increased the luciferase activity in all the cells tested. Adding ER stress to those cells decreased the luciferase activity except in cells silenced with DNAJB12 and DNAJB14.

      These data confirm that DNAJB12 and DNAJB14 are involved in the reflux of ER proteins in general and AGR2 in particular. Inhibition of DNAJB12 and DNAJB14 prevented the inhibitory interaction between AGR2 and wt-p53 and thus rescued wt-p53 phosphorylation and its transcriptional activity as a consequence. “

      Fig3A: overexpression of DNAJB12 decreases Eto induced p53 but not at steady state. Is this because at steady state the activity is already basal? Or is there another reason?

      Answer: yes, at steady state the activity is already basal

      Switch Figs S3D and S3C as they are not referred to in order. Also Fig S3C: vary colour (or add pattern) on bars more between conditions

      Answer: The Figures now are called by their order in the new version. Colors are now added to Figure S3C.

      Need to define HLJ1 at first mention

      Answer: defined as” HLJ1 - High copy Lethal J-protein -an ER-resident tail-anchored HSP40 cochaperone.

      Results section 3

      HSC70 cochaperone (SGTA) defined twice

      Answer: the second one was removed

      "These data are important because SGTA and the ER-resident proteins (PRDX4, AGR2, and DNAJB11) are known to be expressed in different compartments, and the interaction occurs only when those ER-resident proteins localize to the cytosol." Is there a reference for this?

      Answer: Peroxireoxin 4 is the only peroxerodin that is expressed in the ER. AGR2 and DNAJB11 are also ER luminal proteins that are known to be solely expressed in the ER. SGTA is part of the cytosolic quality control system and is expressed in the cytosol. The references are added in the main text.

      Results section 4

      "by almost two folds"

      Answer: corrected

      Fig 6A: It seems strange that the difference between purple and blue bars in scrambled, and D14-KD are very significant but D12-KD is only significant. Why is this? The error bars don't look that different. It would be interesting to see the individual means for each different replicate.

      Answer: Thank you for pointing this, the two asterixis were aligned in the middle as one during figure alignments. In D14 the purple one has a lower error bar thus this changes the significance when compared to the blue while in D12-KD, the error bars in the eto treatment and the eto-Tm both are slightly higher. Graphs of the three different replicates are now added in Figure S6. Each one of the three biological replicates was repeated in three different technical repeats (averaged in the graphs).

      Figures: Fig 6A: Scale bars not well placed. Annotation on final set should be D12/D14 DKD?

      Answer: both were Corrected

      __Discussion __47. The authors mention that they want to use DNAJB12/4-HSC70/SGTA axis to impair cancer cell fitness: What effect would this have though in a non cancer model? Would this be a viable approach Although it is obviously early days, which approach would the authors see as potentially favorable?


      Answer: In our previous study we used an approach to target AGR2 in the cytosol because the reflux of AGR2 occurs only in cancer cells and not in normal cells. In that study we targeted AGR2 with scFv that targets AGR2 and is expressed in the cytosol, in this case it will target AGR2 in the cytosol which only occurs in cancer. Here, we suggest to target the interaction between the refluxed proteins and their new partners in the cytosol or to target the mechanism that causes their reflx to the cytosol by inhibiting for instance the interaction between SGTA and DNAJB proteins.


      __ __ Second para: Should be "Here we present evidences"

      Answer: we replaced with “Here we present evidences”

      "DNAJB12 overexpression was also sufficient to promote ERCYS by refluxing AGR2 and inhibit wt-p53 signaling in cells treated with etoposide" Suggest:

      Answer: DNAJB12 overexpression is also sufficient to promote ERCYS by refluxing AGR2 and inhibit wt-p53 signaling in cancer cells treated with etoposide (Figure 3). This suggests that it is enough to increase the levels of DNAJB12 without inducing the unfolded protein response in order to activate ERCYS. Moreover, the downregulation of DNAJB12 and DNAJB14 rescued the inhibition of wt-p53 during ER stress (Figure 2). Thus, wt-p53 inhibition is independent of the UPR activation but depends on the inhibitory interaction of AGR2 with wt-p53 in the cytosol.

      .

      DNAJB12 overexpression was also sufficient to promote ERCYS by increasing reflux of AGR2 and inhibition of wt-p53 signaling in cells treated with etoposide

      Answer: This sentence is repeated twice and was removed

      "Moreover, DNAJB12 was sufficient to promote this phenomenon and cause ER protein reflux by mass action without causing ER stress (Figure 3, Figure 4, and Figure S3)." You dont look at induction of ER stress here, please change the text or explain in more depth with refs if suitable

      Answer: In the initial submission and in the revised version we assayed the activation of the UPR by looking at the levels of spliced Xbp1 and Bip in the different conditions when DNAJB12 and DNAJB14 are overexpressed (Figure S3C and S3D). Our data show that although DNAJB12 overexpression induces ERCYS, there was no UPR activation.

      The mention of viruses is sparse in this paper. If it is a main theory, put it more centrally to the concept, and explain in more detail. As it is, its appearance in the final sentence is out of context.

      Answer: DNAJB12 and DNAJB14 were reported to facilitate the escape of non-envelope viruses from the endoplasmic reticulum to the cytosol. The mechanism of non-envelope penetration is highly similar to the reflux of proteins from the ER to the cytosol. Interestingly, this mechanism takes place when the DNAJB12 and DNAJB14 form a complex with chaperones from both the ER and the cytosol including HSC70, SGTA and BiP (Walczak et al. 2014; Goodwin et al. 2011; Goodwin et al. 2014)..

      Moreover, the UGGT1 that was independently found in our previous mass spectrometry analysis of the digitonin fraction obtained from HEK293T cells treated with the ER stressor thapsigargin and from isolated human GBM tumors (Sicari et al. 2020), is known to deploy to the cytosol upon viral infection (Huang et al. 2017; Sicari et al. 2020). We therefore hypothesized that the same machinary that is known to allow viruses to escape the ER to penetrate the cytosol may play an important role in the reflux of ER proteins to the cytosol.

      Because ER protein reflux and the penetration of viruses from the ER to the cytosol behave similarly, we speculate that viruses hijacked an evolutionary conserved machinery -ER protein reflux- to penetrate to the cytosol. This is key because it was also reported that during the process of nonenveloped viruses penetration, large, intact and glycosylated viral particles are able to penetrate the ER membrane on their way to the cytosol (Inoue and Tsai 2011).

      Action: we developed the discussion around this point and clarified it better because we believe it central to show that viruses hijacked this conserved mechanism.

      **Referees cross-commenting**

      I agree with the comments from Reviewer 1.

      Reviewer 2 also is correct in many ways, but I think that they have somewhat overlooked the relevance of the ER-stress element and treatments. The authors do need to reference past papers more to give a full story, as this includes the groups own papers, I don't think that it is an ethical problem but rather an oversight in the writing. Regarding reviewer 2's concerns about overexpression levels and cell death, the authors do use an inducible cell line and show the levels of DNAJB12 induced (could CRISPR also be considered?). This could be used to further address reviewer 2's concerns. It would also be useful to see data on cell death in the conditions used in the paper. Re concerns about ER integrity, this could be addressed by using IF (or EM) to show a secondary ER marker that remains ER-localised, and this would also be of interest regarding my comment on which categories of proteins can undergo reflux. If everything is relocalised, then reviewer 2's point would be validated.

      Reviewer #3 (Significance (Required)):

      Significance

      General assessment: This paper robustly shows that the yeast system of ER to cytosol reflux of ER-resident proteins is conserved in mammalian cells, and it describes clearly the link between ER stress, protein reflux and inhibition of p53 in mammalian cells. The authors have the tools to delve deeper into this mechanism and robustly explore this pathway, however the mechanistic elements - where not instantly clear from the results - have been over interpreted somewhat The results have been oversimplified in their explanations and some points and complexities of the study need to be addressed further to make the most of them - these are often some of the more interesting concepts of the paper, for example the differences in DNAJB12/14 and how the proteins orchestrate in the cytosol to play their cytosol-specific effects. I think that many points can be addressed in the text, by the authors being clear and concise with their reporting, while other experiments would turn this paper from an observational one, into a very interesting mechanistic one.

      Advance: This paper is based on previous nice papers from the group. It is a nice progressions from yeast, to basic mechanism, to physiological model. But as mentioned, without a strong mechanistic improvement, the paper would remain observatory.

      Audience: This paper is interesting to cell biologists (homeostasis, quality control and trafficking) as well as cancer cell biologists (fitness of cancer cells and homeostasis) and it is a very interesting demonstration of a process that is a double edged sword, depending on the environment of the cells.

      My expertise: cell biology, trafficking, ER homeostasis

      Answer: We would like to thank the reviewer for his/her positive feedback on our manuscript. All the comments of the three reviewers are now addressed and the manuscript has been strengthen. We put more emphasis on the mechanistic aspect with more Ips and knockdowns. We also added data to show that it is physiologically relevant. We hope that after that the revised version addressed all the concerns raised by the reviewers.

      Goodwin, E. C., A. Lipovsky, T. Inoue, T. G. Magaldi, A. P. Edwards, K. E. Van Goor, A. W. Paton, J. C. Paton, W. J. Atwood, B. Tsai, and D. DiMaio. 2011. 'BiP and multiple DNAJ molecular chaperones in the endoplasmic reticulum are required for efficient simian virus 40 infection', MBio, 2: e00101-11.

      Goodwin, E. C., N. Motamedi, A. Lipovsky, R. Fernandez-Busnadiego, and D. DiMaio. 2014. 'Expression of DNAJB12 or DNAJB14 causes coordinate invasion of the nucleus by membranes associated with a novel nuclear pore structure', PLoS One, 9: e94322.

      Grove, D. E., C. Y. Fan, H. Y. Ren, and D. M. Cyr. 2011. 'The endoplasmic reticulum-associated Hsp40 DNAJB12 and Hsc70 cooperate to facilitate RMA1 E3-dependent degradation of nascent CFTRDeltaF508', Mol Biol Cell, 22: 301-14.

      Huang, P. N., J. R. Jheng, J. J. Arnold, J. R. Wang, C. E. Cameron, and S. R. Shih. 2017. 'UGGT1 enhances enterovirus 71 pathogenicity by promoting viral RNA synthesis and viral replication', PLoS Pathog, 13: e1006375.

      Igbaria, A., P. I. Merksamer, A. Trusina, F. Tilahun, J. R. Johnson, O. Brandman, N. J. Krogan, J. S. Weissman, and F. R. Papa. 2019. 'Chaperone-mediated reflux of secretory proteins to the cytosol during endoplasmic reticulum stress', Proc Natl Acad Sci U S A, 116: 11291-98.

      Inoue, T., and B. Tsai. 2011. 'A large and intact viral particle penetrates the endoplasmic reticulum membrane to reach the cytosol', PLoS Pathog, 7: e1002037.

      Malinverni, D., S. Zamuner, M. E. Rebeaud, A. Barducci, N. B. Nillegoda, and P. De Los Rios. 2023. 'Data-driven large-scale genomic analysis reveals an intricate phylogenetic and functional landscape in J-domain proteins', Proc Natl Acad Sci U S A, 120: e2218217120.

      Piette, B. L., N. Alerasool, Z. Y. Lin, J. Lacoste, M. H. Y. Lam, W. W. Qian, S. Tran, B. Larsen, E. Campos, J. Peng, A. C. Gingras, and M. Taipale. 2021. 'Comprehensive interactome profiling of the human Hsp70 network highlights functional differentiation of J domains', Mol Cell, 81: 2549-65 e8.

      Sicari, D., F. G. Centonze, R. Pineau, P. J. Le Reste, L. Negroni, S. Chat, M. A. Mohtar, D. Thomas, R. Gillet, T. Hupp, E. Chevet, and A. Igbaria. 2021. 'Reflux of Endoplasmic Reticulum proteins to the cytosol inactivates tumor suppressors', EMBO Rep: e51412.

      Sicari, Daria, Raphael Pineau, Pierre-Jean Le Reste, Luc Negroni, Sophie Chat, Aiman Mohtar, Daniel Thomas, Reynald Gillet, Ted Hupp, Eric Chevet, and Aeid Igbaria. 2020. 'Reflux of Endoplasmic Reticulum proteins to the cytosol yields inactivation of tumor suppressors', bioRxiv.

      Walczak, C. P., M. S. Ravindran, T. Inoue, and B. Tsai. 2014. 'A cytosolic chaperone complexes with dynamic membrane J-proteins and mobilizes a nonenveloped virus out of the endoplasmic reticulum', PLoS Pathog, 10: e1004007.

      Yamamoto, Y. H., T. Kimura, S. Momohara, M. Takeuchi, T. Tani, Y. Kimata, H. Kadokura, and K. Kohno. 2010. 'A novel ER J-protein DNAJB12 accelerates ER-associated degradation of membrane proteins including CFTR', Cell Struct Funct, 35: 107-16.

      Youker, R. T., P. Walsh, T. Beilharz, T. Lithgow, and J. L. Brodsky. 2004. 'Distinct roles for the Hsp40 and Hsp90 molecular chaperones during cystic fibrosis transmembrane conductance regulator degradation in yeast', Mol Biol Cell, 15: 4787-97.

    1. My grandfather is Italian, so my dad is half Italian. Can I call ...Quorahttps://www.quora.com › My-grandfather-is-Italian-so-...Quorahttps://www.quora.com › My-grandfather-is-Italian-so-...

      just kinda like au-jus ...

      i was of course taught that if yo momma not a ju then you not a ju unless you subscribe to the idea that if the temple won't open its arms and doors to you its

      just like a busted ass chew chew chew chew see seeeee we are the r c wayyyyyyyy veeeee .......

      ADOM

      ARCHMIA

      THEADOM

      when will i solidify around the "vestibule of autho...

      RISE

      seattle ... .... ... ... .. .. .. wa wa wa waw wa wa wwa wa wa w

      "the best of the est" honestly i'm pretty sure its gonna iwnd up being nejmac and lax but if it's not that then it might just be the people i don't really know must have something to do with

      what is the diference between ECOLA ECOLA

      AGRICOLA ERAT "I.N.R. ... why?" what's a Villae ?? Now MCO is helping us find our "oh ya we no ...|"

      honestly is it MO? is it DLP? is it keep it on the "u" and the "n" and maybe you'll find the creation of disclose d is cl closer to Reno and St. Louis

      before you find out about Lewis and C the Ark wade through the .... idea that we have

      MOTORCADE BARRICADE

      we have hints in words that there was once a timeline called "A.D." and if you don't rmemeber that then you are BCE

    1. "Am I really understanding from that other person's point of view?" That commitment to understanding the argument from the other person's perspective. Now, what sympathy in this case means is not that I feel what they feel, it's that I'm willing to set aside, even if it's just temporarily, that hunt for the slightest misstep in logic or reasoning; setting that aside for a moment so that I can listen really carefully to what my conversation partner is saying, so I can understand, from their perspective, what their intellectual project is, or why it is that they are looking at the same world I'm looking at, but coming to a very, very different conclusion.
    1. upport cross-divisional thinking and that the best ideas are already in a company and it's just a matter of sort of um getting people together

      for - neuroscience - validation for Stop Reset Go open source participatory system mapping for design innovation

      neuroscience - validation for Stop Reset Go open source participatory system mapping for design innovation - bottom-up collective design efficacy - What Henning Beck validates for companies can also apply to using Stop Reset Go participatory system mapping within an open space to de-silo and be as inclusive as possible of many different silo'd transition actors

      • Presenter discusses improvements in PowerPoint skills and the development of a new implementation of Idris.

        • "I've been working very hard on my PowerPoint skills recently this is a this is a kind of a show-and-tell session so I've been hiding away in my cupboard over the last few months working on a new implementation of address."
      • Curious about attendees' experience with dependently typed programming languages.

        • "Has anyone played with any dependently type of programming languages before just by way of show of hands?"
      • Overview of Idris improvements, focusing on type system advancements.

        • "Firstly definitely faster type checking...I don't get out much you know so them so I'm gonna show you a bit about the new interactive additive feature."
      • Emphasis on interactive editing and user-friendly type-driven development.

        • "It's all about the interactive editing it's all about the condo that the machine and the program of being in a conversation."
      • Introduction of quantitative type theory and its significance in Idris.

        • "The key new feature of the language though is that it's based on quantitative type theory."
      • Explanation of quantities in types: 0, 1, and many, and their runtime implications.

        • "0 means this is going to be a race at runtime and the type checker has told me that it's going to be raised at runtime 1 means I can use this exactly once at runtime and many is just back in the old world where we always used to be."
      • Demonstration of the use of quantities in practical examples like vectors.

        • "The basic idea is the lengths are encoded in the type...so if I'm writing an append function and I say the effector has n things and another vector has n things then the resulting vector has n plus M States."
      • Exploration of automated programming with interactive editing, including a magic trick analogy.

        • "Wouldn't it be nice if we get abstract away in the acts of programming so if we could just say you know keep case splitting and searching until you find an answer and hilariously that works quite well."
      • Advantages of having more information in types for program synthesis.

        • "The linearity gives more information to the type prototype directed synthesis."
      • Implementation of session types using linear and dependent types.

        • "With linearity and dependent types, you absolutely can do that and I think it's quite a nice way of writing concurrent programs."
      • Encouragement to contribute to Idris development on GitHub.

        • "Please come along I'd love to have your funky issues right so this isn't dependent type stock so by law I have to show you the vexes there's a reason I'm going to show you the vectors."
      • Final thoughts on interactive editing, efficiency, and community contribution.

        • "Interactive editing there's a lot more we can do with interactive editing really important that it's responsive so this is I want this to be the default mode in which people interact with the system if that's going to be the case it has to come back with the answer very quickly."
    1. Written inPython, Cython

      Is this accurate? I don't have a lot of firsthand experience with data science stuff, but usually when looking just past surface-level you find that some Python package is really a shell around some native binary core implemented in e.g. C (or Fortran?).

      When I at the repos for spaCy and its assistant Thinc, GitHub's language analysis shows that it's pretty much Python. Is there something lurking in the shadows that I'm not seeing? Or does this mean that if someone cloned spaCy and Thinc and wrote it in JS, then the subset of data scientists whose work can be done with those two packages (and whatever datavis generators they use) will benefit from the faster runtime and the the elimination of figging and other setup?

    1. Welcome back. In this lesson, I want to introduce another core AWS service, the simple storage service known as S3. If you use AWS in production, you need to understand S3. This lesson will give you the very basics because I'll be deep diving into a specific S3 section later in the course, and the product will feature constantly as we go. Pretty much every other AWS service has some kind of interaction with S3. So let's jump in and get started.

      S3 is a global storage platform. It's global because it runs from all of the AWS regions and can be accessed from anywhere with an internet connection. It's a public service. It's regional based because your data is stored in a specific AWS region at rest. So when it's not being used, it's stored in a specific region. And it never leaves that region unless you explicitly configure it to. S3 is regionally resilient, meaning the data is replicated across availability zones in that region. S3 can tolerate the failure of an AZ, and it also has some ability to replicate data between regions, but more on that in the S3 section of the course.

      Now S3 might initially appear confusing. If you utilize it from the UI, you appear not to have to select a region. Instead, you select the region when you create things inside S3, which I'll talk about soon. S3 is a public service, so it can be accessed from anywhere as long as you have an internet connection. The service itself runs from the AWS public zone. It can cope with unlimited data amounts and it's designed for multi-user usage of that data. So millions of users could be accessing cute cat pictures added by the Animals for Life Rescue Officers. S3 is perfect for hosting large amounts of data. So think movies or audio distribution, large scale photo storage like stock images, large textual data or big data sets. It could be just as easily used for millions or billions of IOT devices or to store images for a blog. It scales from nothing to near unlimited levels.

      Now S3 is economical, it's a great value service for storing and allowing access to data. And it can be accessed using a variety of methods. There's the GUI, you can use the command line, the AWS APIs or even standard methods such as HTTP. I want you to think of S3 as the default storage service in AWS. It should be your default starting point unless your requirement isn't delivered by S3. And I'll talk more about the limitations and use cases later in this lesson.

      S3 has two main things that it delivers: Objects and Buckets. Objects are the data the S3 stores, your cat picture, the latest episode of Game of Thrones, which you have stored legally, of course, or it could be large scale datasets showing the migration of the koala population in Australia after a major bushfire. Buckets are containers for objects. It's buckets and objects that I want to cover in this lesson as an introduction to the service.

      So first, let's talk about objects. An object in S3 is made up of two main components and some associated metadata. First, there is the object key. And for now you can think of the object key, similar to a file name. The key identifies the object in a bucket. So if you know the object key and the bucket, then you can uniquely access the object, assuming that you have permissions. Remember by default, even for public services, there is no access in AWS initially, except for the account root user.

      Now, the other main component of an object is its value. And the value is the data or the contents of the object. In this case, a sequence of binary data, which represents a koala in his house. In this course, I want to avoid suggesting that you remember pointless values. Sometimes though, there are things that you do need to commit to memory. And this is one of those times. The value of an object, in essence, how large the object is, can range from zero bytes up to five terabytes in size. So you can have an empty object or you can have one that is a huge five TB. This is one of the reasons why S3 is so scalable and so useful in a wide range of situations because of this range of sizes for objects.

      Now, the other components of an object, aren't that important to know at this stage, but just so you hear the terms that I'll use later, objects also have a version ID, metadata, some access control, as well as sub resources. Now don't worry about what they do for now, I'll be covering them all later. For this lesson, just try to commit to memory what an object is, what components it has and the size range for an object.

      So now that we've talked about objects, let's move on and look at buckets. Buckets are created in a specific AWS region. And let's use Sydney or ap-southeast-2 as an example. This has two main impacts. Firstly, your data that's inside a bucket has a primary home region. And it never leaves that region, unless you as an architect or one of your system admins configures that data to leave that region. That means that S3 has stable and controlled data sovereignty. By creating a bucket in a region, you can control what laws and rules apply to that data. What it also means is that the blast radius of a failure is that region.

      Now this might be a new term. What I mean by blast radius is that if a major failure occurs, say a natural disaster or a large scale data corruption, the effect of that will be contained within the region. Now a bucket is identified by its name, the bucket name in this case, koala data. A bucket name needs to be globally unique. So that's across all regions and all accounts of AWS. If I pick a bucket name, in this case, koala data, nobody else can use it in any AWS account. Now making a point of stressing this as it often comes up in the exam. Most AWS things are often unique in a region or unique in your account. For example, I can have an IAM user called Fred and you can also have an IAM user called Fred. Buckets though are different, with buckets, the name has to be totally unique, and that's across all regions and all AWS accounts. I've seen it come up in the exam a few times. So this is definitely a point to remember.

      Now buckets can hold an unlimited number of objects. And because objects can range from zero to five TB in size, that essentially means that a bucket can hold from zero to unlimited bytes of data. It's an infinitely scalable storage system. Now one of the most important things that I want to say in this lesson is that as an object storage system, an S3 bucket has no complex structure. It's flat, it has flat structure. All objects stored within the bucket at the same level. So this isn't like a file system where you can truly have files within folders, within folders. Everything is stored in the bucket at the root level.

      But, if you do a listing on an S3 bucket, you will see what you think are folders. Even the UI presents this as folders. But it is important for you to know at this stage that that's not how it actually is. Imagine a bucket where you see three image files, koala one, two and three dot JPEG. The first thing is that inside S3, there's no concept of file type based on the name. These are just three objects where the object key is koala1.JPEG, koala2.JPEG and koala3.JPEG. Now folders in S3 are represented when we have object names that are structured like these. So the objects have a key, a forward slash old forward slash koala one, two and three dot JPEG. When we create object names like this, then S3 presents them in the UI as a folder called old. So because we've got object names that begin with slash old, then S3 presents this as a folder called old. And then inside that folder, we've got koala one, two, and three dot JPEG.

      Now nine out of 10 times, this detail doesn't matter, but I want to make sure that you understand how it actually works. Folders are often referred to as prefixes in S3 because they're part of the object names. They prefix the object names. As you move through the various stages of your AWS learnings, this is gonna make more and more sense. And I'm gonna demonstrate this in the next lesson, which is a demo lesson.

      Now to summarize buckets are just containers, they're stored in a region, and for S3, they're generally where a lot of permissions and options are set. So remember that buckets are generally the default place where you should go to, to configure the way the S3 works.

      Now, I want to cover a few summary items and then step through some patterns and anti-patterns for S3, before we move to the demo. But first an exam powerup. These are things that you should try to remember and they will really help in the exam. First bucket names are globally unique. Remember that one because it will really help in the exam. I've seen a lot of times where AWS have included trick questions, which test your knowledge of this one. If you get any errors, you can't create a bucket a lot of the time it's because somebody else already has that bucket name.

      Now bucket names do have some restrictions. They need to be between 3 and 63 characters, all lower case and no underscores. They have to start with a lowercase letter or a number, and they can't be formatted like IP addresses. So you can't have 1.1.1.1 as your bucket name. Now there are some limits in terms of buckets. Now limits are often things that you don't need to remember for the exam, but this is one of the things that you do. There is a limit of a hundred buckets that you can have in an AWS account. So this is not per region, it's for the entire account. There's a soft limit of 100 and a hard limit so you can increase all the way up to this hard limit using support requests, and this hard limit is a thousand.

      Now this matters for architectural reasons. It's not just an arbitrary number. If you're designing a system which uses S3 and users of that system store data inside S3, you can implement a solution that has one bucket per user if you have anywhere near this number of users. So if you have anywhere from a hundred to a thousand users or more of a system, then you're not gonna be able to have one bucket per user because you'll hit this hard limit. You tend to find this in the exam quite often, it'll ask you how to structure a system, which has potentially thousands of users. What you can do is take a single bucket and divide it up using prefixes, so those folders that aren't really folders, and then in that way, you can have multiple users using one bucket. Remember the 100/1000, it's a 100 soft limit and a 1000 hard limit.

      You aren't limited in terms of objects in a bucket, you can have zero to an infinite number of objects in a bucket. And each object can range in size from zero bytes to 5 TB in size. And then finally, in terms of the object structure, an object consists of a key, which is its name and then the value, which is the data. And there are other elements to an object which I'll discuss later in the course, but for now, just remember the two main components, the key and the value. Now, all of these points are worth noting down, maybe make them into a set of flashcards and you can use them later on during your studies.

      S3 is pretty straightforward and that there tend to be a number of things that it's really good at and a fairly small set of things that it's not suitable for. So let's take a look. S3 is an object storage system. It's not a file storage system, and it's not a block storage system, which are the other main types. What this means is that if you have a requirement where you're accessing the whole of these entities, so the whole of an object, so an image, an audio file, and you're doing all of that at once, then it's a candidate for object storage. If you have a Window server which needs to access a network file system, then it's not S3 that needs to be file-based storage. S3 has no file system, it's flat. So you can't browse to an S3 bucket like you would a file share in Windows. Likewise, it's not block storage, which means you can't mount it as a mount point or a volume on the Linux or Windows. When you're dealing with virtual machines or instances, you mount block storage to them. Block storage is basically virtual hard disks. In EC2, you have EBS, which is block storage. Block storage is generally limited to one thing accessing it at a time, one instance in the case of EBS. S3 doesn't have that single user limitation and it's not block storage, but that means you can't mount it as a drive.

      S3 is great for large scale data storage or distribution. Many examples I'll show you throughout the course will fit into that category. And it's also good for offloading things. If you have a blog with lots of posts and lots of images or audio or movies, instead of storing that data on an expensive compute instance, you can move it to an S3 bucket and configure your blog software to point your users at S3 directly. You can often shrink your instance by offloading data onto S3. And don't worry, I'll be demoing this later in the course. Finally, S3 should be your default thought for any input to AWS services or output from AWS services. Most services which consume data and or output data can have S3 as an option to take data from or put data to when it's finished. So if you're faced with any exam questions and there's a number of options on where to store data, S3 should be your default. There are plenty of AWS services which can output large quantities of data or ingest large quantities of data. And most of the time, it's S3, which is an ideal storage platform for that service.

      Okay time for a quick demo. And in this demo, we're just gonna run through the process of creating a simple S3 bucket, uploading some objects to that bucket, and demonstrating exactly how the folder functionality works inside S3. And I'm also gonna demonstrate a number of elements of how access and permissions work with S3. So go ahead and complete this video, and when you're ready join me in the next, which is gonna be a demo of S3.

    1. Reviewer #1 (Public Review):

      Summary:

      The authors aimed to develop a mean-field model that captures the key aspects of activity in the striatal microcircuit of the basal ganglia. They start from a spiking network of individual neuron models tuned to fit striatal data. They show that an existing mean-field framework matches the output firing rates generated by the spiking network both in static conditions and when the network is subject to perfectly periodic drive. They introduce a very simplified representation of dopaminergic cortico-striatal plasticity and show that simulated dopamine exposure makes model firing rates go up or down, in a way that matches the design of the model. Finally, they aim to test the performance of the model in a reinforcement learning scenario, with two very simplified channels corresponding to the selection between two actions. Overall, I do not find that this work will be useful for the field or provide novel insights.

      Strengths:

      The mean-field model dynamics match well with the spiking network dynamics in all scenarios shown. The authors also introduce a dopamine-dependent synaptic plasticity rule in the context of their reinforcement learning task, which can nicely capture the appropriate potentiation or depression of corticostriatal synapses when dopamine levels change.

      Weaknesses:

      From the title onwards, the authors refer to a "multiscale" model. They do not, in fact, work with a multiscale model; rather, they fit a spiking model to baseline data and then fit a mean-field model to the spiking model. The idea is then to use the mean-field model for subsequent simulations.

      The mean-field modeling framework that is used was already introduced previously by the authors, so that is not a novel aspect of this work in itself. The model includes an adaptation variable for each population in the network. Mean-field models with adaptation already exist, and there is no discussion of why this new framework would be preferable to those. Moreover, as presented, the mean-field model is not a closed system. It includes a variable w (in equation 7) that is never defined.

      Overall, the paper shows that a mean-field model behaves similarly to a spiking model in several scenarios. A much stronger result would be to show that the mean-field model captures the activity of neurons recorded experimentally. The spiking model is supposedly fit to data from recordings in some sort of baseline conditions initially, but the quality of this fit is not adequately demonstrated; the authors just show a cursory comparison of data from a single dSPN neuron with the activity of a single model dSPN, for one set of parameters.

      The authors purport to test their model via its response to "the main brain rhythms observed experimentally". In reality, this test consists of driving the model with periodic input signals. This is far too simplistic to achieve the authors' goals in this part of the work.

      The work also presents model responses to simple simulations of dopamine currents, treated as negative or positive inputs to different model striatal populations. These are implemented as changes in glutamate conductance and possibly in an additional depolarizing/hyperpolarizing current, so the results that are shown are guaranteed to occur by the direct design of the simulation experiment; nothing new is learned from this. The consideration of dopamine also points out that the model is apparently designed and fit in a way that does not explicitly include dopamine, even though the fitting is done to control (i.e., with-dopamine) data, so it's not clear how this modeling framework should be adapted for dopamine-depleted scenarios.

      For the reinforcement learning scenario, the model network considered is extremely simplified. Moreover, the behavior generated is unrealistic, with action two selected several times in succession independent of reward outcomes and then an instant change to a pattern of perfectly alternating selection of action 1 and action 2.

      Finally, various aspects of the paper are sloppily written. The Discussion section is especially disappointing, because it is almost entirely a summary of the results of the paper, without an actual discussion of their deeper implications, connections to the existing literature, predictions that emerge, caveats or limitations of the current work, and natural directions for future study, as one would expect from a usual discussion section.

    1. Note: This response was posted by the corresponding author to Review Commons. The content has not been altered except for formatting.

      Learn more at Review Commons


      Reply to the reviewers

      Reviewer 1

      R1 Cell profiling is an emerging field with many applications in academia and industry. Finding better representations for heterogeneous cell populations is important and timely. However, unless convinced otherwise after a rebuttal/revision, the contribution of this paper, in our opinion, is mostly conceptual, but in its current form - not yet practical. This manuscript combined two concepts that were previously reported in the context of cell profiling, weakly supervised representations. Our expertise is in computational biology, and specifically applications of machine learning in microscopy.

      In our revised manuscript, we have aimed to better clarify the practical contributions of our work by demonstrating the effectiveness of the proposed concepts on real-world datasets. We hope that these revisions and our detailed responses address your concerns and highlight the potential impact of our approach.

      R1.1a. CytoSummaryNet is evaluated in comparison to aggregate-average profiling, although previous work has already reported representations that capture heterogeneity and self-supervision independently. To argue that both components of contrastive learning and sets representations are contributing to MoA prediction we believe that a separate evaluation for each component is required. Specifically, the authors can benchmark their previous work to directly evaluate a simpler population representation (PMID: 31064985, ref #13) - we are aware that the authors report a 20% improvement, but this was reported on a separate dataset. The authors can also compare to contrastive learning-based representations that rely on the aggregate (average) profile to assess and quantify the contribution of the sets representation.

      We agree that evaluating the individual contributions of the contrastive learning framework and single-cell data usage is important for understanding CytoSummaryNet's performance gains.

      To assess the impact of the contrastive formulation independently, we applied CytoSummaryNet to averaged profiles from the cpg0004 dataset. This isolated the effect of contrastive learning by eliminating single-cell heterogeneity. The experiment yielded a 32% relative improvement in mechanism of action retrieval, compared to the 68% gain achieved with single-cell data. These findings suggest that while the contrastive formulation contributes significantly to CytoSummaryNet's performance, leveraging single-cell information is crucial for maximizing its effectiveness. We have added a discussion of this experiment to the Results section:

      “We conducted an experiment to determine whether the improvements in mechanism of action retrieval were due solely to CytoSummaryNet's contrastive formulation or also influenced by the incorporation of single-cell data. We applied the CytoSummaryNet framework to pre-processed average profiles from the 10 μM dose point data of Batch 1 (cpg0004 dataset). This approach isolated the effect of the contrastive architecture by eliminating single-cell data variability. We adjusted the experimental setup by reducing the learning rate by a factor of 100, acknowledging the reduced task complexity. All other parameters remained as described in earlier experiments.

      This method yielded a less pronounced but still substantial improvement in mechanism of action retrieval, with an increase of 0.010 (32% enhancement - Table 1). However, this improvement was not as high as when the model processed single-cell level data (68% as noted above). These findings suggest that while CytoSummaryNet's contrastive formulation contributes to performance improvements, the integration of single-cell data plays a critical role in maximizing the efficacy of mechanism of action retrieval.”

      We don't believe comparing with PMID: 31064985 is useful: while the study showcased the usefulness of modeling heterogeneity using second-order statistics, its methodology is limited in scalability due to the computational burden of computing pairwise similarities for all perturbations, particularly in large datasets. Additionally, the study's reliance on similarity network fusion, while expedient, introduces complexity and inefficiency. We contend that this comparison does not align with our objective of testing the effectiveness of heterogeneity in isolation, as it primarily focuses on capturing second and first-order information. Thus, we do not consider this study a suitable baseline for comparison.

      R1.1b. The evaluation metric of mAP improvement in percentage is misleading, because a tiny improvement for a MoA prediction can lead to huge improvement in percentage, while a much larger improvement in MoA prediction can lead to a small improvement in percentage. For example, in Fig. 4, MEK inhibitor mAP improvement of ~0.35 is measured as ~50% improvement, while a much smaller mAP improvement can have the same effect near the origins (i.e., very poor MoA prediction).

      We agree that relying solely on percentage improvements can be misleading, especially when small absolute changes result in large percentage differences.

      However, we would like to clarify two key points regarding our reporting of percentage improvements:

      • We calculate the percentage improvement by first computing the average mAP across all compounds for both CytoSummaryNet and average profiling, and then comparing these averages. This approach is less susceptible to the influence of outlier improvements compared to calculating the average of individual compound percentage improvements.
      • We report percentage improvements alongside their corresponding absolute improvements. For example, the mAP improvement for Stain4 (test set) is reported as 0.052 (60%). To further clarify this point, we have updated the caption of Table 1 to explicitly state how the percentage improvements are calculated:

      The improvements are calculated as mAP(CytoSummaryNet)-mAP(average profiling). The percentage improvements are calculated as (mAP(CytoSummaryNet)-mAP(average profiling))/mAP(average profiling).

      R1.1b. (Subjective) visual assessment of this figure does not show a convincing contribution of CytoSummaryNet representations of the average profiling on the test set (3.33 uM). This issue might also be relevant for the task of replicate retrieval. All in all, the mAP improvement reported in Table 1 and throughout the manuscript (including the Abstract), is not a proper evaluation metric for CytoSummaryNet contribution. We suggest reporting the following evaluations:

      1. Visualizing the results of cpg0001 (Figs. 1-3) similarly to cpg0004 (Fig. 4), i.e., plotting the matched mAP for CytoSummaryNet vs. average profile.

      2. In Table 1, we suggest referring to the change in the number of predictable MoAs (MoAs that pass a mAP threshold) rather than the improvement in percentages. Another option is showing a graph of the predictability, with the X axis representing a threshold and Y-axis showing the number of MoAs passing it. For example see (PMID: 36344834, Fig. 2B) and (PMID: 37031208, Fig. 2A), both papers included contributions from the corresponding author of this manuscript.

      Regarding the suggestion to visualize the results for compound group cpg0001 similarly to cpg0004, unfortunately, this is not feasible due to the differences in data splitting between the two datasets. In cpg0001, an MoA might have one compound in the training set and another in the test or validation set. Reporting a single value per MoA would require combining these splits, which could be misleading as it would conflate performance across different data subsets.

      However, we appreciate the suggestion to represent the number of predictable MoAs that surpass a certain mAP threshold, as it provides another intuitive measure of performance. To address this, we have created a graph that visualizes the predictability of MoAs across various thresholds, similar to the examples provided in the referenced papers (PMID: 36344834, Figure 2B and PMID: 37031208, Figure 2A). This graph, with the x-axis depicting the threshold and the y-axis showing the number of MoAs meeting the criterion, has been added to Supplementary Material K.

      R1.1c.i. "a subset of 18 compounds were designated as validation compounds" - 5 cross-validations of 18 compounds can make the evaluation complete. This can also enhance statistical power in figures 1-3.

      We appreciate your suggestion and acknowledge the potential benefits of employing cross-validation, particularly in enhancing statistical power. While we understand the merit of cross-validation for evaluating model performance and generalization to unseen data, we believe the results as presented already highlight the generalization characterics of our methods.

      Specifically, (the new) Figure 3 demonstrates the model's improvement over average profiling in both training and validation plates, supporting its ability to generalize to unseen compounds (but not to unseen plates).

      While cross-validation could potentially enhance our analysis, retraining five new models solely for different validation set results may not substantially alter our conclusions, given the observed trends in Suppl Figure A1 and (the new) Figure 4, both of which show results across multiple stain sets (but a single train-test-validation split).


      R1.1c.ii. Clarify if the MoA results for cpg0001 are drawn from compounds from both the training and the validation datasets. If so, describe how the results differ between the sets in text and graphs.

      We confirm that the Mechanism of Action (MoA) retrieval results for cpg0001 are derived from all available compounds. It's important to note that the training and validation dataset split for the replicate retrieval task is different from the MoA prediction task. For replicate retrieval, we train using all available compounds and validate on a held-out set (see Figure 2). For MoA prediction, we train using the replicate retrieval task as the objective on all available compounds but validate using MoA retrieval, which is a distinct task. We have added a brief clarification in the main text to highlight the distinction between these tasks and how validation is performed for each:

      “We next addressed a more challenging task: predicting the mechanism of action class for each compound at the individual well level, rather than simply matching replicates of the exact same compound (Figure 5). It's also important to note that mechanism of action matching is a downstream task on which CytoSummaryNet is not explicitly trained. Consequently, improvements observed on the training and validation plates are more meaningful in this context, unlike in the previous task where only improvements on the test plate were meaningful. For similar reasons, we calculate the mechanism of action retrieval performance on all available compounds, combining both the training and validation sets. This approach is acceptable because we calculate the score on so-called "sister compounds" only—that is, different compounds that have the same mechanism of action annotation. This ensures there is no overlap between the mechanism of action retrieval task and the training task, maintaining the integrity of our evaluation. ”

      R1.1c.iii. "Mechanism of action retrieval is evaluated by quantifying a profile's ability to retrieve the profile of other compounds with the same annotated mechanism of action.". It was unclear to us if the evaluation of mAP for MoA identification can include finding replicates of the same compound. That is, whether finding a close replicate of the same compound would be included in the AP calculation. This would provide CytoSummaryNet with an inherent advantage as this is the task it is trained to do. We assume that this was not the case (and thus should be more clearly articulated), but if it was - results need to be re-evaluated excluding same-compound replicates.

      The evaluation excludes replicate wells of the same compound and only considers wells of other compounds with the same MoA. This methodology ensures that the model's performance on the MoA prediction task is not inflated by its ability to find replicates of the same compound, which is the objective of the replicate retrieval task. Please see the explanation we have added to the main text in our response to R1.1c.ii. Additionally, we have updated the Methods section to clearly describe this evaluation procedure:

      “Mechanism of action retrieval is evaluated by quantifying a profile’s ability to retrieve the profile of different compounds with the same annotated mechanism of action.”



      __R1.2a. __The description of Stain2-5 was not clear for us at first (and second) read. The information is there, but more details will greatly enhance the reader's ability to follow. One suggestion is explicitly stating that these "stains" partitioning was already defined in ref 26. Another suggestion is laying out explicitly a concrete example on the differences between two of these stains. We believe highlighting the differences between stains will strengthen the claim of the paper, emphasizing the difficulty of generalizing to the out-of-distribution stain.

      We appreciate your feedback on the clarity of the Stain2-5 dataset descriptions; we certainly struggled to balance detail and concepts in describing these. We have made the following changes:

      • Explicitly mentioned that the partitioning of the Stain experiments was defined in https://pubmed.ncbi.nlm.nih.gov/37344608/: “The partitioning of the Stain experiments have been defined and explained previously [21].”
      • Moved an improved version of (now) Figure 2 from the Methods section to the main text to help visually explain how the stratification is done early on.
      • Added a new section in the Experimental Setup: Diversity of stain sets, which includes a concrete example highlighting the differences between Stain2, and Stain5 to emphasize the diversity in experimental setups within the same dataset: “Stain2-5 comprise a series of experiments which were conducted sequentially to optimize the experimental conditions for image-based cell profiling. These experiments gradually converged on the most optimal set of conditions; however, within each experiment, there were significant variations in the assay across plates. To illustrate the diversity in experimental setups within the dataset, we will highlight the differences between Stain2 and Stain5.

      Stain2 encompasses a wide range of nine different experimental protocols, employing various imaging techniques such as Widefield and Confocal microscopy, as well as specialized conditions like multiplane imaging and specific stains like MitoTracker Orange. This subset also includes plates acquired with strong pixel binning instead of default imaging and plates with varying concentrations of dyes like Hoechst. As a result, Stain2 exhibits greater variance in the experimental conditions across different plates compared to Stain5.

      In contrast, Stain5, the last experiment in the series, follows a more systematic approach, consistently using either confocal or default imaging across three well-defined conditions. Each condition in Stain5 utilizes a lower cell density of 1,000 cells per well compared to Stain2's 2,500 cells per well. Being the final experiment in the series, Stain5 had the least variance in experimental conditions.

      For training the models, we typically select the data containing the most variance to capture the broadest range of experimental variation. Therefore, we chose Stain2-4 for training, as they represented the majority of the data and captured the most experimental variation. We reserved Stain5 for testing to evaluate the model's ability to generalize to new experimental conditions with less variance.

      All StainX experiments were acquired in different passes, which may introduce additional batch effects.”

      These changes aim to provide a clearer understanding of the dataset's complexity and the challenges associated with generalizing to out-of-distribution data.

      R1.2b. What does each data point in Figures 1-3 represent? Is it the average mAP for the 18 validation compounds, using different seeds for model training? Why not visualize the data similarly to Fig. 4 so the improvement per compound can be clearly seen?

      The data points in (the new) Figures 3,4,5 represent the average mAP for each plate, calculated by first computing the mAP for each compound and then averaging across compounds to obtain the average mAP per plate. We have updated the figure captions to clarify this:

      "... (each data point is the average mAP of a plate) ..."

      While visualizing the mAP per compound, similar to (the new) Figure 6 for cpg0004, could provide insights into compound-level improvements, it would require creating numerous additional figures or one complex figure to adequately represent all the stratifications we are analyzing (plate, compound, Stain subset). By averaging the data per plate across different stratifications, we aim to provide a clearer and more comprehensible overview of the trends and improvements while allowing us to draw conclusions about generalization.

      Please note: this comment is related to the comment R1.1b (Subjective)

      R1.2.c [On the topic of enhancing clarity and readability:] Justification and interpretation of the evaluation metrics.

      Please refer to our response to comment R1.1b, where we have addressed your concerns regarding the justification and interpretation of the evaluation metrics.

      R1.2d. Explicitly mentioning the number of MoAs for each datasets and statistics of number of compounds per MoA (e.g., average\median, min, max).

      We have added the following to the Experimental Setup: Data section:

      “A subset of the data was used for evaluating the mechanism of action retrieval task, focusing exclusively on compounds that belong to the same mechanism class. The Stain plates contained 47 unique mechanisms of action, with each compound replicated four times. Four mechanisms had only a single compound; the four mechanisms (and corresponding compounds) were excluded, resulting in 43 unique mechanisms used for evaluation. In the LINCS dataset, there were 1436 different mechanisms, but only 661 were used for evaluation because the remaining had only one compound.”

      R1.2e. The data split in general is not easily understood. Figure 8 is somewhat helpful, however in our view, it can be improved to enhance understanding of the different splits. Specifically, the training and validation compounds need to be embedded and highlighted within the figure.

      Thank you for highlighting this. We have completely revised the figure, now Figure 2 which we hope more clearly conveys the data split strategy.

      Please note: this comment is related to the comment R1.2a.





      R1.3a. Why was stain 5 used for the test, rather than the other stains?

      Stain2-5 were part of a series of experiments aimed at optimizing the experimental conditions for image-based cell profiling using Cell Painting. These experiments were conducted sequentially, gradually converging on the most optimal set of conditions. However, within each experiment, there were significant variations in the assay across plates, with earlier iterations (Stain2-4) having more variance in the experimental conditions compared to Stain5. As Stain5 was the last experiment in the series and consisted of only three different conditions, it had the least variance. For training the models, we typically select the data containing the most variance to capture the broadest range of experimental variation. Therefore, Stain2-4 were chosen for training, while Stain5 was reserved for testing to evaluate the model's ability to generalize to new experimental conditions with less variance.

      We have now clarified this in the Experimental Setup: Diversity of stain sets section. Please see our response to comment R1.2a. for the full citation.

      R1.3b How were the 18 validation compounds selected?

      20% of the compounds (n=18) were randomly selected and designated as validation compounds, with the remaining compounds assigned to the training set. We have now clarified this in the Results section:

      “Additionally, 20% of the compounds (n=18) were randomly selected and designated as validation compounds, with the remaining compounds assigned to the training set (Supplementary Material H).”

      R1.3c. For cpg0004, no justification for the specific doses selected (10uM - train, 3.33 uM - test) for the analysis in Figure 4. Why was the data split for the two dosages? For example, why not perform 5-fold cross validation on the compounds (e.g., of the highest dose)?

      We chose to use the 10 μM dose point as the training set because we expected this higher dosage to consist of stronger profiles with more variance than lower dose points, making it more suitable for training a model. We decided to use a separate test set at a different dose (3.33 μM) to assess the model's ability to generalize to new dosages. While cross-validation on the highest dose could also be informative, our approach aimed to balance the evaluation of the model's generalization capability with its ability to capture biologically relevant patterns across different dosages.

      This explanation has been added to the text:

      “We chose the 10 μM dose point for training because we expected this high dosage to produce stronger profiles with more variance than lower dose points, making it more suitable for model training.”

      “The multiple dose points in this dataset allowed us to create a separate hold-out test set using the 3.33 μM dose point data. This approach aimed to evaluate the model's performance on data with potentially weaker profiles and less variance, providing insights into its robustness and ability to capture biologically relevant patterns across dosages. While cross-validation on the 10 μM dose could also be informative, focusing on lower dose points offers a more challenging test of the model's capacity to generalize beyond its training conditions, although we do note that all compounds’ phenotypes would likely have been present in the 10 μM training dataset, given the compounds tested are the same in both.”

      R1.3d. A more detailed explanation on the logic behind using a training stain to test MoA retrieval will help readers appreciate these results. In our first read of this manuscript we did not grasp that, we did in a second read, but spoon-feeding your readers will help.

      This comment is related to the rationale behind training on one task and testing on another, which is addressed in our responses to comments R1.1.cii and R1.1.ciii.

      R1.4 Assessment of interpretability is always tricky. But in this case, the authors can directly confirm their interpretation that the CytoSummaryNet representation prioritizes large uncrowded cells, by explicitly selecting these cells, and using their average profile re

      We progressively filtered out cells based on a quantile threshold for Cells_AreaShape features (MeanRadius, MaximumRadius, MedianRadius, and Area), which were identified as important in our interpretability analysis, and then computed average profiles using the remaining cells before determining the replicate retrieval mAP. In the exclusion experiment, we gradually left out cells as the threshold increased, while in the inclusion experiment, we progressively included larger cells from left to right.

      The results show that using only the largest cells does not significantly increase the performance. Instead, it is more important to include the large cells rather than only including small cells. The mAP saturates after a threshold of around 0.4, indicating that larger cells define the profile the most, and once enough cells are included to outweigh the smaller cell features, the profile does not change significantly by including even larger cells.

      These findings support our interpretation that CytoSummaryNet prioritizes large, uncrowded cells. While this approach could potentially be used as a general outlier removal strategy for cell profiling, further investigation is needed to assess its robustness and generalizability across different datasets and experimental conditions.

      We have created Supplementary Material L to report these findings and we additionally highlight them in the Results:

      “To further validate CytoSummaryNet's prioritization of large, uncrowded cells, we progressively filtered cells based on Cells_AreaShape features and observed the impact on replicate retrieval mAP (Supplementary Material L). The results support our interpretation and highlight the key role of larger cells in profile strength.”

      __R1.5. __Placing this work in context of other weakly supervised representations. Previous papers used weakly supervised labels of proteins / experimental perturbations (e.g., compounds) to improve image-derived representations, but were not discussed in this context. These include PMID: 35879608, https://www.biorxiv.org/content/10.1101/2022.08.12.503783v2 (from the same research groups and can also be benchmarked in this context), https://pubs.rsc.org/en/content/articlelanding/2023/dd/d3dd00060e , and https://www.biorxiv.org/content/10.1101/2023.02.24.529975v1. We believe that a discussion explicitly referencing these papers in this specific context is important.

      While these studies provide valuable insights into improving cell population profiles using representation learning, our work focuses specifically on the question of single-cell aggregation methods. We chose to use classical features for our comparisons because they are the current standard in the field. This approach allows us to directly assess the performance of our method in the context of the most widely used feature extraction pipeline in practice. However, we see the value in incorporating them in future work and have mentioned them in the Discussion:

      “Recent studies exploring image-derived representations using self-supervised and self-supervised learning [35][36] could inspire future research on using learned embeddings instead of classical features to enhance model-aggregated profiles.”

      R1.minor1. "Because the improved results could stem from prioritizing certain features over others during aggregation, we investigated each cell's importance during CytoSummaryNet aggregation by calculating a relevance score for each" - what is the relevance score? Would be helpful to provide some intuition in the Results.

      We have included more explanation of the relevance score in the Results section, following the explanation of sensitivity analysis (SA) and critical point analysis (CPA):

      “SA evaluates the model's predictions by analyzing the partial derivatives in a localized context, while CPA identifies the input cells with the most significant contribution to the model's output. The relevance scores of SA and CPA are min-max normalized per well and then combined by addition. The combination of the two is again min-max normalized, resulting in the SA and CPA combined relevance score (see Methods for details).”

      R1.minor2. Figure 1:

      1. Colors of the two methods too similar
      2. The dots are too close. It will be more easily interpreted if they were further apart.
      3. What do the dots stand for?
      4. We recommend considering moving this figure to the supp. material (the most important part of it is the results on the test set and it appears in Fig.2).
      1. We chose a lighter and darker version of the same color as a theme to simplify visualization, as this theme is used throughout (the new) Figures 3,4,5.
      2. We agree; we have now redrawn the figure to fix this.
      3. Each data point is the average mAP of a plate. Please see our answer for R1.2b as well.
      4. We believe that (the new) Figures 3,4,5 serve distinct purposes in testing various generalization hypotheses. We have added the following text to emphasize that the first figures are specifically about generalization hypothesis testing: “We first investigated CytoSummaryNet’s capacity to generalize to out-of-distribution data: unseen compounds, unseen experimental protocols, and unseen batches. The results of these investigations are visualized in Figures 3, 4, and 5, respectively.”

      R1.minor3 Figure 4: It is somewhat misleading to look at the training MoAs and validation MoAs embedded together in the same graph. We recommend showing only the test MoAs (train MoAs can move to SI).

      We addressed this comment in R1.1c.ii. To reiterate briefly, there are no training, validation, or test MoAs because these are not used as labels during the training process. There is an option to split them based on training and validation compounds, which is addressed in R1.1c.ii.


      R1.minor4 Figure 5

      1. Why only Stain3? What happens if we look at Stains 2,3 and 4 together? Stain 5?

      2. Should validation compounds and training compounds be analyzed separately?

      3. Subfigure (d): it is expected that the data will be classified by compound labels as it is the training task, but for this to be persuasive I would like to see this separately on the training compounds first and then and more importantly on the validation compounds.

      4. For subfigures (b) and (d): it appears there are not enough colors for d, which makes it partially not understandable. For example, the pink label in (d) shows a single compound which appears to represent two different MoAs. This is probably not the case, and it has two different compounds, but it cannot be inferred when they are represented by the same color.

      5. For the Subfigure (e) - only 1 circle looks justified (in the top left). And for that one, is it not a case of an outlier plate that would perhaps need to be removed from analysis? Is it not good that such a plate will be identified?

      We have addressed this point in the text, stating that the results are similar for Stain2 and Stain4. Stain5 represents an out-of-distribution subset because of a very different set of experimental conditions (see Experimental Setup: Diversity of stain sets). To improve clarity, we have revised the figure caption to reiterate this information:

      “... Stain2 and Stain4 yielded similar results (data not shown). …”

      1. For replicate retrieval, analyzing validation and training compounds separately is appropriate. However, this is not the case for MoA retrieval, as discussed in our responses to R1.1c.ii and R1.1c.i.
      2. We have created the requested plot (below) but ultimately decided not to include it in the manuscript because we believe that (the new) Figures 3 and 4 are more effective for making quantitative comparative claims.

      [Please see the full revision document for the figures]

      Top: training compounds (validation compounds grayed out); not all compounds are listed in the legend.

      *Bottom: validation compounds (training compounds grayed out). *

      Left: average profiling; Right: CytoSummaryNet

      1. We agree with your observation and have addressed this issue by labeling the center mass as a single class (gray) and highlighting only the outstanding pairs in color. Please refer to the updated figure and our response to R3.6 for more details.

      2. In the updated figure, we have revised the figure caption to focus solely on the annotation of same mechanism of action profile clusters, as indicated by the green ellipses. The annotation of isolated plate clusters has been removed (Figures 7e and 7f) to maintain consistency and avoid potential confusion. Despite being an outlier for Stain3, the plate (BR00115134bin1) clusters with Stain4 plates (Supplementary Figure F1, green annotated square inside the yellow annotated square), indicating it is not merely a noisy outlier and can provide insights into the out-of-sample performance of our model.

      R1.minor5a. Discussion: "perhaps in part due to its correction of batch effects" - is this statement based on Fig. 5F - we are not convinced.

      We appreciate the reviewer's scrutiny regarding our statement about batch effect correction. Upon reevaluation, we agree that this claim was not adequately substantiated by empirical data. We quantified the batch effects using comparison mean average precision for both average profiles and CytoSummaryNet profiles, and the statistical analysis revealed no significant difference between these profiles in terms of batch effect correction. Therefore, we have removed this theoretical argument from the manuscript entirely to ensure that all claims are strongly supported by the data presented.

      R1.minor5b. "Overall, these results improve upon the ~20% gains we previously observed using covariance features" - this is not the same dataset so it is hard to reach conclusions - perhaps compare performance directly on the same data?

      We have now explicitly clarified this is a different dataset. Please see our response to R1.1a for why a direct comparison was not performed. The following clarification can be found in the Discussion:

      “These results improve upon the ~20% gains previously observed using covariance features [13] albeit on a different dataset, and importantly, CytoSummaryNet effectively overcomes the challenge of recomputation after training, making it easier to use.”

      Reviewer 2

      R2.1 The authors present a well-developed and useful algorithm. The technical motivation and validation are very carefully and clearly explained, and their work is potentially useful to a varied audience.

      That said, I think the authors could do a better job, especially in the figures, of putting the algorithm in context for an audience that is unfamiliar with the cell painting assay. (a) For example, a figure towards the beginning of the paper with example images might help to set the stage. (b) Similarly a schematic of the algorithm earlier in the paper would provide a graphical overview. (c) For the sake of a biologically inclined audience, I would consider labeling the images in the caption by cell type and label.

      Thank you for your valuable suggestions on improving the accessibility of our figures for readers unfamiliar with the Cell Painting assay. We have made the following changes to address your comments:

      1. and b. To provide visual context and a graphical overview of the algorithm, we have moved the original Figure 7 to Figure 1. This figure now includes example images that help readers new to the Cell Painting assay.
      2. We have added relevant details to the example images in (the new) Figure 1

        R2.2 The interpretability results were intriguing. The authors might consider further validating these interpretations by removing weakly informative cells from the dataset and retraining. Are the cells so uninformative that the algorithm does better without them, or are they just less informative than other cells?

      Please see our responses to R1.4 and R3.0

      R2.3 As far as I can tell, the authors only oblique state whether the code associated with the manuscript is openly available. Posting the code is needed for reproducibility. I would provide not only a github, but a doi linked to the code, or some other permanent link.

      We have now added a Code Availability and Data Availability section, clearing stating that the code and data associated with the manuscript are openly available.

      R2.4 Incorporating biological heterogeneity into machine-learning driven problems is a critical research question. Replacing means/modes and such with a machine learning framework, the authors have identified a problem with potentially wide significance. The application to cell painting and related assays is of broad enough significance for many journals, However, the authors could further broaden the significance by commenting on other possible cell biology applications. What other applications might the algorithm be particularly suited for? Are there any possible roadblocks to wider use. What sorts of data has the code been tested on so far?

      We have added the following paragraph to discuss the broader applicability of CytoSummaryNet:

      “The architecture of CytoSummaryNet holds significant potential for broader applications beyond image-based cell profiling, accommodating tabular, permutation-invariant data and enhancing downstream task performance when applied to processed population-level profiles. Its versatility makes it valuable for any omics measurements where downstream tasks depend on measuring similarity between profiles. Future research could also explore CytoSummaryNet's applicability to genetic perturbations, expanding its utility in functional genomics.”

      Reviewer 3

      R3.0 The authors have done a commendable job discussing the method, demonstrating its potential to outperform current models in profiling cell-based features. The work is of considerable significance and interest to a wide field of researchers working on the understanding of cell heterogeneity's impact on various biological phenomena and practical studies in pharmacology.

      One aspect that would further enhance the value of this work is an exploration of the method's separation power across different modes of action. For instance, it would be interesting to ascertain if the method's performance varies when dealing with actions that primarily affect size, those that affect marker expression, or compounds that significantly diminish cell numbers.

      Thank you for encouraging comments!

      We have added the following to Results: Relevance scores reveal CytoSummaryNet's preference for large, isolated cells:

      “Statistical t-tests were conducted to identify the features that most effectively differentiate mechanisms of action from negative controls in average profiles, focusing on the three mechanisms of action where CytoSummaryNet demonstrates the most significant improvement and the three mechanisms where it shows the least. Consistent with our hypothesis that CytoSummaryNet emphasizes larger, more sparse cells, the important features for the CytoSummaryNet-improved mechanisms of action (Supplementary Material I1) often involve the radial distribution for the mitochondria and RNA channels. These metrics capture the fraction of those stains near the edge of the cell versus concentric rings towards the nucleus, which are more readily detectable in larger cells compared to small, rounded cells.

      In contrast, the important features for mechanisms of action not improved by CytoSummaryNet (Supplementary Material I) predominantly include correlation metrics between brightfield and various fluorescent channels, capturing spatial relationships between cellular components. Some of these mechanisms of action included compounds that were not individually distinguishable from negative controls, and CytoSummaryNet did not overcome the lack of phenotype in these cases. This suggests that while CytoSummaryNet excels in identifying certain cellular features, its effectiveness is limited when dealing with mechanisms of action that do not exhibit pronounced phenotypic changes.”

      We have also added supplementary material to support (I. Relevant features for CytoSummaryNet improvement).

      R3.0 Another test on datasets that are not concerned with chemical compounds, but rather genetic perturbations would greatly increase the reach of the method into the functional genomics community and beyond. This additional analysis could provide valuable insights into the versatility and applicability of the proposed method.

      We agree that testing the method's behavior on genetic perturbations would be interesting and could provide insights into its versatility. However, the efficacy of the methodology may vary depending on the specific properties of different genetic perturbation types.

      For example, the penetrance of phenotypes may differ between genetic and chemical perturbations. In some experimental setups, a selection agent ensures that nearly all cells receive a genetic perturbation (though not all may express a phenotype due to heterogeneity or varying levels of the target protein). Other experiments may omit such an agent. Additionally, different patterns might be observed in various classes of reagents, such as overexpression, CRISPR-Cas9 knockdown (CRISPRn), CRISPR-interference (CRISPRi), and CRISPR-activation (CRISPRa).

      We believe that selecting a single experiment with one of these technologies would not adequately address the question of versatility. Instead, we propose future studies that may conclusively assess the method's performance across a variety of genetic perturbation types. This would provide a more comprehensive understanding of CytoSummaryNet's applicability in functional genomics and beyond. We have update the Discussion section to reflect this:

      “Future research could also explore CytoSummaryNet's applicability to genetic perturbations, expanding its utility in functional genomics.”

      R3.1. The datasets were stratified based on plates and compounds. It would be beneficial to clarify the basis for data stratification applied for compounds. Was the data sampled based on structural or functional similarity of compounds? If not, what can be expected from the model if trained and validated using structurally or functionally diverse and non-diverse compounds?

      Thank you for raising the important question of data stratification based on compound similarity. In our study, the data stratification was performed by randomly sampling the compounds, without considering their structural or functional similarity.

      This approach may limit the generalizability of the learned representations to new structural or functional classes not captured in the training set. Consequently, the current methodology may not fully characterize the model’s performance across diverse compound structures.

      In future work, it would be valuable to explore the impact of compound diversity on model performance by stratifying data based on structural or functional similarity and comparing the results to our current random stratification approach to more thoroughly characterize the learned representations.

      R3.2. Is the method prioritizing a particular biological reaction of cells toward common chemical compounds, such as mitotic failure? Could this be oncology-specific, or is there more utility to it in other datasets?

      Our analysis of CytoSummaryNet's performance in (the new) Figure 6 reveals a strong improvement in MoAs targeting cancer-related pathways, such as MEK, HSP, MDM, dehydrogenase, and purine antagonist inhibitors. These MoAs share a common focus on cellular proliferation, survival, and metabolic processes, which are key characteristics of cancer cells.

      Given the composition of the cpg0004 dataset, which contains 1,258 unique MoAs with only 28 annotated as oncology-related, the likelihood of randomly selecting five oncology-related MoAs that show strong improvement is extremely low. This suggests that the observed prioritization is not due to chance.

      Furthermore, the prioritization cannot be solely attributed to the frequency of oncology-related MoAs in the dataset. Other prevalent disease areas, such as neurology/psychiatry, infectious disease, and cardiology, do not exhibit similar improvements despite having higher MoA counts.

      While these findings indicate a potential prioritization of oncology-related MoAs by CytoSummaryNet, further research is necessary to fully understand the extent and implications of this bias. Future work should involve conducting similar analyses across other disease areas and cell types to assess the method's broader utility and identify areas for refinement and application. However, given the speculative nature of these observations, we have chosen not to update the manuscript to discuss this potential bias at this time.

      R3.3 Figures 1 and 2 demonstrate that the CytoSummaryNet profiles outperform average-aggregated profiles. However, the average profiling results seem more consistent when compared to CytoSummaryNet profiling. What further conditions or approaches can help improve CytoSummaryNet profiling results to be more consistent?

      The observed variability in CytoSummaryNet's performance is primarily due to the intentional technical variance in our datasets, where each plate tested different staining protocol variations. It's important to note that this level of technical variance is not typical in standard cell profiling experiments. In practice, the variance across plates would be much lower. We want to emphasize that while a model capable of generalizing across diverse experimental conditions might seem ideal, it may not be as practically useful in real-world scenarios. This is because such non-uniform conditions are uncommon in typical cell profiling experiments. In normal experimental settings, where technical variance is more controlled, we expect CytoSummaryNet's performance to be more consistent.

      R3.4 Can the poor performance on unseen data (in the case of stain 5) be overcome? If yes, how? If no, why not?

      We believe that the poor performance on unseen data, such as Stain 5, can be overcome depending on the nature of the unseen data. As shown in Figure 4 (panel 3), the model improves upon average profiling for unseen data when the experimental conditions are similar to the training set.

      The issue lies in the different experimental conditions. As explained in our response to R3.3, this could be addressed by including these experimental conditions in the training dataset. As long as CytoSummaryNet is trained (seen) and tested (unseen) on data generated under similar experimental conditions, we are confident that it will improve or perform as well as average profiling.

      It's important to note that the issue of generalization to vastly different experimental conditions was considered out of scope for this paper. The main focus is to introduce a new method that improves upon average profiling and can be readily used within a consistent experimental setup.

      R3.5 It needs to be mentioned how the feature data used for CytoSummaryNet profiling was normalized before training the model. What would be the impact of feature data normalization before model training? Would the model still outperform if the skewed feature data is normalized using square or log transformation before model training?

      We have clarified in the manuscript that we standardized the feature data on a plate-by-plate basis to achieve zero mean and unit variance across all cells per feature within each plate. We have added the following statement to improve clarity:

      “The data used to compute the average profiles and train the model were standardized at the plate-level, ensuring that all cell features across the plate had a zero mean and unit variance. The negative control wells were then removed from all plates."

      We chose standardization over transformations like squaring or logging to maintain a balanced scale across features while preserving the biological and morphological information inherent in the data. While transformations can reduce skewness and are useful for data spanning several orders of magnitude, they might distort biological relevance by compressing or expanding data ranges in ways that could obscure important cellular variations.

      Regarding the potential impact of square or log transformations on skewed feature data, these methods could improve the model's learning efficiency by making the feature distribution more symmetrical. However, the suitability and effectiveness of these techniques would depend on the specific data characteristics and the model architecture.

      Although not explored in this study, investigating various normalization techniques could be a valuable direction for future research to assess their impact on the performance and adaptability of CytoSummaryNet across diverse datasets and experimental setups.

      R3.6. In Figure 5 b and c, MoAs often seem to be represented by singular compounds and thus, the test (MoA prediction) is very similar to the training (compound ID). Given this context, a discussion about the extent this presents a circular argument supported by stats on the compound library used for training and testing would be beneficial.

      Clusters in (the new) Figure 7 that contain only replicates of a single compound would not yield an improved performance on the MoA task unless they also include replicates of other compounds sharing the same MoA in close proximity. Please see our response to R1.1c.iii. for details. To improve visual clarity and avoid misinterpretation, we have recomputed the colors for (the new) Figure 7 and grayed out overlapping points.

      R3.7 Can you estimate the minimum amount of supervision (fuzzy/sparse labels, often present in mislabeled compound libraries with dirty compounds and polypharmacology being present) that is needed for it to be efficiently trained?

      It's important to note that the metadata used by the model is only based on identifying replicates of the same compound. Mechanism of action (MoA) annotations, which can be erroneous due to dirty compounds, polypharmacology, and incomplete information, are not used in training at all. MoA annotations are only used in our evaluation, specifically for calculating the mAP for MoA retrieval.

      We have successfully trained CytoSummaryNet on 72 unique compounds with 4 replicates each. This is the current empirical minimum, but it is possible that the model could be trained effectively with even fewer compounds or replicates.

      Determining the absolute minimum amount of supervision required for efficient training would require further experimentation and analysis. Factors such as data quality, feature dimensionality, and model complexity could influence the required level of supervision.

      R3.minor1 Figure 5: The x-axis and y-axis tick values are too small, and image resolution/size needs to be increased.

      We have made the following changes to address the concerns:

      • Increased the image resolution and size to improve clarity and readability.
      • Removed the x-axis and y-axis tick values, as they do not provide meaningful information in the context of UMAP visualizations. We believe these modifications enhance the visual presentation of the data and make it easier for readers to interpret the results.

      R3.minor2 The methods applied to optimize hyperparameters in supplementary data need to be included.

      We added the following to Supplementary Material D:

      “We used the Weights & Biases (WandB) sweep suite in combination with the BOHB (Bayesian Optimization and HyperBand) algorithm for hyperparameter sweeps. The BOHB algorithm [47] combines Bayesian optimization with bandit-based strategies to efficiently find optimal hyperparameters.

      Additionally Table D1 provides an overview of all tunable hyperparameters and their chosen values based on a BOHB hyperparameter optimization.”

      R3.minor3 Figure 5(c, d): The names of compound 2 and Compound 5 need to be included in the labels.

      These compounds were obtained from external companies and are proprietary, necessitating their anonymization in our study. This has now been added in the caption of (the new) Figure 7:

      “Note that Compound2 and Compound5 are intentionally anonymized.”

      R3.minor4 Table C1: Plate descriptions need to be included.

      *Table C1: The training, validation, and test set stratification for Stain2, Stain3, Stain4, and Stain5. Five training, four validation, and three test plates are used for Stain2, Stain3, and Stain4. Stain5 contains six test set plates only. *

      __Stain2 __

      Stain3

      Stain4

      Stain5

      Training plates

      Test plates

      BR00113818

      BR00115128

      BR00116627

      BR00120532

      BR00113820

      BR00115125highexp

      BR00116631

      BR00120270

      BR00112202

      BR00115133highexp

      BR00116625

      BR00120536

      BR00112197binned

      BR00115131

      BR00116630highexp

      BR00120530

      BR00112198

      BR00115134

      200922_015124-Vhighexp

      BR00120526

      Validation plates

      BR00120274

      BR00112197standard

      BR00115129

      BR00116628highexp

      BR00112197repeat

      BR00115133

      BR00116629highexp

      BR00112204

      BR00115128highexp

      BR00116627highexp

      BR00112201

      BR00115127

      BR00116629

      Test plates

      BR00112199

      BR00115134bin1

      200922_044247-Vbin1

      BR00113819

      BR00115134multiplane

      200922_015124-V

      BR00113821

      BR00115126highexp

      BR00116633bin1

      We have added a reference to the metadata file in the description of Table C1: https://github.com/carpenter-singh-lab/2023_Cimini_NatureProtocols/blob/main/JUMPExperimentMasterTable.csv

      R3.minor5 Figure F1: Does the green box (stain 3) also involve training on plates from stain 4 (BR00116630highexp) and 5 (BR00120530) mentioned in Table C1? Please check the figure once again for possible errors.

      We have carefully re-examined Figure F1 and Table C1 to ensure their accuracy and consistency. Upon double-checking, we can confirm that the figure is indeed correct. We intentionally omitted the training and validation plates from Figure F1 to maintain clarity and readability, as including them resulted in a cluttered and difficult-to-interpret figure.

      Regarding the specific plates mentioned:

      • BR00116630highexp (Stain4) is used for training, as correctly stated in Table C1. This plate is considered an outlier within the Stain4 dataset and happens to cluster with the Stain3 plates in Figure F1.
      • BR00120530 (Stain5) is part of the test set only and correctly falls within the Stain5 cluster in Figure F1. To improve the clarity of the training, validation, and test split in Table C1, we have added a color scheme that visually distinguishes the different data subsets. This should make it easier for readers to understand the distribution of plates across the various splits.
    1. Welcome back. In this lesson, I want to cover the architecture of public AWS services and private AWS services. This is foundational to how AWS works, from a networking and security perspective. The differences might seem tiny, but understanding them fully will help you grasp more complex network and security products or architectures throughout your studies.

      AWS services can be categorized into two main types: public services and private services. If you don’t have much AWS experience, you might assume that a public service is accessible to everyone, and a private service isn't. However, when you hear the terms AWS private service and AWS public service, it’s referring to networking only. A public service is accessed using public endpoints, such as S3, which can be accessed from anywhere with an internet connection. A private AWS service runs within a VPC, so only things within that VPC, or connected to that VPC, can access the service. For both, there are permissions as well as networking. Even though S3 is a public service, by default, an identity other than the account root user has no authorization to access that resource. So, permissions and networking are two different considerations when talking about access to a service. For this lesson, it's the networking which matters.

      When thinking about any sort of public cloud environment, most people instinctively think of two parts: the internet and private network. The internet is where internet-based services operate, like online stores, Gmail, and online games. If you're at home playing an online game or watching training videos, you’re connecting to the internet via an internet service provider. So this is the internet zone. Then we have the private network. If you’re watching this video from home, your home network is an example of a private network. Only things directly connected to a network port within your house or people with your WiFi password can operate in your personal, private network zone.

      AWS also has private zones called Virtual Private Clouds (VPCs). These are isolated, so VPCs can't communicate with each other unless you allow it, and nothing from the internet can reach these private networks unless you configure it. Services like EC2 instances can be placed into these private zones and, just like with your home network, it can only access the internet, and the internet can only access it if you allow and configure it.

      Many people think AWS is architected with just two network zones: the internet and private zones. But there's actually a third zone: the AWS public zone, which runs between the public internet and the AWS private zone networks. This is not on the public internet but connected to it. The distinction might seem irrelevant, but it matters as you learn more about advanced AWS networking. The AWS public zone is where AWS public services operate, like S3.

      To summarize, there are three different network zones: the public internet, the AWS private zone (where VPCs run), and the AWS public zone (where AWS public services operate). If you access AWS public services from anywhere with a public internet connection, your communication uses the public internet for transit to and from this AWS public zone. This is why you can access AWS public services from anywhere with an internet connection because the internet is used to carry packets from you to the AWS public zone and back again.

      Later in the course, I will cover how you can configure virtual or physical connections between on-premises networks and AWS VPCs, allowing private networks to connect together if you allow it. You can also create and attach an internet gateway to a VPC, allowing private-zone resources to access the public internet if they have a public IP address. This also allows access to public AWS services like S3 without touching the public internet, communicating through the AWS public zone.

      Private resources, such as EC2 instances, can be given a public IP address, allowing them to be accessed from the public internet. Architecturally, this projects the EC2 instance into the public zone, enabling communication with the public internet. Understanding the three different network zones—the public internet, the AWS public zone, and the AWS private zone—is crucial for doing well in the real world and in specific exams. These three network zones become critical as you learn more advanced networking features of AWS.

      That’s everything for this lesson. Complete the video, and when you’re ready, I’ll look forward to you joining me in the next lesson.

    1. Reviewer #1 (Public Review):

      The study uses nanoscale secondary ion mass spectrometry to show that maize plants inoculated with a bacteria, Gd, incorporated fixed nitrogen into the chloroplast. The authors then state that since "chloroplasts are the chief engines that drive plant growth," that it is this incorporation that explains the maize's enhanced growth with the bacteria.

      But the authors don't present the total special distribution of nitrogen in plants. That is, if the majority of nitrogen is in the chloroplast (which, because of Rubisco, it likely is) then the majority of fixed nitrogen should go into the chloroplast.

      Also, what are the actual controls? In the methods, the authors detail that the plants inoculated with Gd are grown without nitrogen. But how did the authors document the "enhanced growth rates of the plants containing this nitrogen fixing bacteria." Were there other plants grown without nitrogen and the Gd? If so, of course, they didn't grow as well. Nitrogen is essential for plant growth. If Gd isn't there to provide it in n-free media, then the plants won't grow. Do we need to go into the mechanism for this, really? And it's not just because nitrogen is needed in the chloroplast, even if that might be where the majority ends up.

      Furthermore, it is not novel to say that nitrogen from a nitrogen fixing bacteria makes its way into the chloroplast. For any plant ever successfully grown on N free media with a nitrogen fixing bacteria, this must be the case. We don't need a fancy tool to know this.

      The experimental setup does not suit the argument the authors are trying to make (and I'm not sure if the argument the authors are trying to make has any legitimacy). The authors contend that their study provides the basis of a "detailed agronomic analysis of the extent of fixed nitrogen fertilizer needs and growth responses in autonomous nitrogen-fixing maize plants." But what is a "fixed nitrogen fertilizer need"? The phrase makes no sense. A plant has nitrogen needs. This nitrogen can be provided via nitrogen fixing bacteria or fertilizer. But are there fixed nitrogen fertilizer needs? It sounds like the authors are suggesting that a plant can distinguish between nitrogen fixed by bacteria nearby and that provided by fertilizer. If that is the contention, then a new set of experiments is needed - with other controls grown on different levels of fertilizer.

      What is interesting, and potentially novel, in this study is figure 1D (and lines 90-99). In that image, is the bacteria actually in the plant cell? Or is it colonizing the region between the cells? Either way, it looks to have made its way into the plant leaf, correct? I believe that would be a novel and fascinating finding. If the authors were to go into more detail into how Gd is entering into the symbiotic relationship with maize (e.g. fixing atmospheric nitrogen in the leaf tissue rather than in root nodules like legumes) I believe that would be very significant. But be sure to add to the field in relation to reference 9, and any new references since then.

      Also, it would be helpful to have an idea of how fast these plants, grown in n free media but inoculated with the bacteria, grow compared to plants grown on various levels of fertilizer.

  3. Jul 2024
    1. Cohort studies increasingly collect biosamples for molecular profiling and are observing molecular heterogeneity. High throughput RNA sequencing is providing large datasets capable of reflecting disease mechanisms. Clustering approaches have produced a number of tools to help dissect complex heterogeneous datasets, however, selecting the appropriate method and parameters to perform exploratory clustering analysis of transcriptomic data requires deep understanding of machine learning and extensive computational experimentation. Tools that assist with such decisions without prior field knowledge are nonexistent. To address this we have developed Omada, a suite of tools aiming to automate these processes and make robust unsupervised clustering of transcriptomic data more accessible through automated machine learning based functions. The efficiency of each tool was tested with five datasets characterised by different expression signal strengths to capture a wide spectrum of RNA expression datasets. Our toolkit’s decisions reflected the real number of stable partitions in datasets where the subgroups are discernible. Within datasets with less clear biological distinctions, our tools either formed stable subgroups with different expression profiles and robust clinical associations or revealed signs of problematic data such as biased measurements.

      This work has been peer reviewed in GigaScience (see paper), which carries out open, named peer-review. These reviews are published under a CC-BY 4.0 license and were as follows:

      Reviewer name: **Pierre Cauchy **

      Kariotis et al present Omada, a tool dedicated to automated partitioning of large-scale, cohort-based RNA-Sequencing data such as TCGA. A great strength for the manuscript is that it clearly shows that Omada is capable of performing partitioning from PanCan into BRCA, COAD and LUAD (Fig 5), and datasets with no known groups (PAH and GUSTO), which is impressive and novel. I would like to praise the authors for coming up with such a tool, as the lack of a systematic tool dedicated to partitioning TCGA-like expression data is indeed a shortcoming in the field of medical genomics Overall, I believe the tool will be very valuable to the scientific community and could potentially contribute to meta-analysis of cohort RNA-Seq data. I only have a few comments regarding the methodology and manuscript. I also think that it should be more clearly stated that Omada is dedicated to large datasets (e.g. TCGA) and not differential expression analysis. I would also suggest benchmarking Omada to comparable tools via ROC curves if possible (see below). Methods: This section should be a bit more homogeneous between text descriptive and mathematical descriptive. It should specify what parts are automated and what part needs user input and refer to the vignette documentation. I also could not find the Omada github repository. Sample and gene expression preprocessing: To me, this section lacks methods/guidelines and only loosely describes the steps involved. "numerical data may need to be normalised in order to account for potential misdirecting quantities" - which kind of normalisation? "As for the number of genes, it is advised for larger genesets (>1000 genes) to filter down to the most variable ones before the application of any function as genes that do not vary across samples do not contribute towards identifying heterogeneity" What filtering is recommended? Top 5% variance? 1%? Based on what metric? Determining clustering potential: To me, it was not clear if this is automatically performed by Omada and how the feasibility score is determined. Intra-method Clustering Agreement: Is this from normalised data? Because affinity matrix will be greatly affected whether it's normalised or non-normalised data as the matrix of exponential(-normalised gene distance)^2 Spectral clustering step 2: "Define D to be the diagonal matrix whose (i, i)-element is the sum of A's i-th row": please also specify that A(i,j) is 0 in this diagonal matrix. Please also confirm which matrix multiplication method is used, product or Cartesian product? Also if there are 0 values, NAs will be obtained in this step. Hierarchical clustering step 5: "Repeat Step 3 a total of n − 1 times until there is only one cluster left." This is a valuable addition as this merges identical clusters, the methods should emphasise that the benefits of this clustering reduction method to help partition data, i.e. that this minimises the number of redundant clusters. Stability-based assessment of feature sets: "For each dataset we generate the bootstrap stability for every k within range". Here it should be mentioned that this is carried out by clusterboot, and the full arguments should be given for documentation "The genes that comprise the dataset with the highest stability are the ones that compose the most appropriate set for the downstream analysis" - is this the single highest or a gene list in the top n datasets? Please specify. Choosing k number of clusters: "This approach prevents any bias from specific metrics and frees the user from making decisions on any specific metric and assumptions on the optimal number of clusters.". Out of consistency with the cluster reduction method in the "intra-clustering agreement" section which I believe is a novelty introduced by Omada, and within the context of automated analysis, the package should also ideally have an optimized number of k-clusters. K-means clustering analysis is often hindered due to the output often resulting in redundant, practically identical clusters which often requires manual merging. While I do understand the rationale described there and in Table 3, in terms of biological information and especially for deregulated genes analysis (e.g. row z-score clustering), should maximum k also not be determined by the number of conditions, i.e 2n, e.g. when n=2, kmax=4; n=3, kmax=8? Test datasets and Fig 6: Please expand on how the number of features 300 was determined. While this number of genes corresponds to a high stability index, is this number fixed or can it be dynamically estimated from a selection (e.g. from 100 to 1000)? Results Overall this section is well written and informative. I would just add the following if applicable: Figure 3: I think this figure could additionally include benchmarking, ROC curves of. Omada vs e.g. previous TCGA clustering analyses (PMID 31805048) Figure 4: I think it would be useful to compare Omada results to previous TCGA clustering analyses, e.g. PMID 35664309 Figure 6: swap C and D. Why is cluster 5 missing on D)?

    1. Author response:

      We thank you for the opportunity to provide a concise response. The criticisms are accurately summarized in the eLife assessment:

      the study fails to engage prior literature that has extensively examined the impact of variance in offspring number, implying that some of the paradoxes presented might be resolved within existing frameworks.

      The essence of our study is to propose the adoption of the Haldane model of genetic drift, based on the branching process, in lieu of the Wright-Fisher (WF) model, based on sampling, usually binomial.  In addition to some extensions of the Haldane model, we present 4 paradoxes that cannot be resolved by the WF model. The reviews suggest that some of the paradoxes could be resolved by the WF model, if we engage prior literature sufficiently.

      We certainly could not review all the literature on genetic drift as there must be thousands of them. Nevertheless, the literature we do not cover is based on the WF model, which has the general properties that all modifications of the WF model share.  (We should note that all such modifications share the sampling aspect of the WF model. To model such sampling, N is imposed from outside of the model, rather than self-generating within the model.  Most important, these modifications are mathematically valid but biologically untenable, as will be elaborated below. Thus, in concept, the WF and Haldane models are fundamentally different.)

      In short, our proposal is general with the key point that the WF model cannot resolve these (and many other) paradoxes.  The reviewers disagree (apparently only partially) and we shall be specific in our response below.

      We shall first present the 4th paradox, which is about multi-copy gene systems (such as rRNA genes and viruses, see the companion paper). Viruses evolve both within and between hosts. In both stages, there are severe bottlenecks.  How does one address the genetic drift in viral evolution? How can we model the effective population sizes both within- and between- hosts?  The inability of the WF model in dealing with such multi-copy gene systems may explain the difficulties in accounting for the SARS-CoV-2 evolution. Given the small number of virions transmitted between hosts, drift is strong which we have shown by using the Haldane model (Ruan, Luo, et al. 2021; Ruan, Wen, et al. 2021; Hou, et al. 2023). 

      As the reviewers suggest, it is possible to modify the WF model to account for some of these paradoxes. However, the modifications are often mathematically convenient but biologically dubious. Much of the debate is about the progeny number, K.  (We shall use haploid model for this purpose but diploidy does not pose a problem as stated in the main text.) The modifications relax the constraint of V(k) = E(k) inherent in the WF sampling.  One would then ask how V(k) can be different from E(k) in the WF sampling even though it is mathematically feasible (but biologically dubious)?  Kimura and Crow (1963) may be the first to offer a biological explanation.  If one reads it carefully, Kimura's modification is to make the WF model like the Haldane model. Then, why don't we use the Haldane model in the first place by having two parameters, E(k) and V(k), instead of the one-parameter WF model?

      The Haldane model is conceptually simpler. It allows the variation in population size, N, to be generated from within the model, rather than artificially imposed from outside of the model.  This brings us to the first paradox, the density-dependent Haldane model. When N is increasing exponentially as in bacterial or yeast cultures, there is almost no drift when N is very low and drift becomes intense as N grows to near the carrying capacity.  We do not see how the WF model can resolve this paradox, which can otherwise be resolved by the Haldane model.

      The second and third paradoxes are about how much mathematical models of population genetic can be detached from biological mechanisms. The second paradox about sex chromosomes is rooted in the realization of V(k) ≠ E(k).  Since E(k) is the same between sexes but V(k) is different, how does the WF sampling give rise to V(k) ≠ E(k)? We are asking a biological question that troubled Kimura and Crow (1963) alluded above. The third paradox is acknowledged by two reviewers. Genetic drift manifested in the fixation probability of an advantageous mutation is 2s/V(k).  It is thus strange that the fundamental parameter of drift in the WF model, N (or Ne), is missing.  In the Haldane model, drift is determined by V(k) with N being a scaling factor; hence 2s/V(k) makes perfect biological sense,

      We now answer the obvious question: If the model is fundamentally about the Haldane model, why do we call it the WF-Haldane model? The reason is that most results obtained by the WF model are pretty good approximations and the branching process may not need to constantly re-derive the results.  At least, one can use the WF results to see how well they fit into the Haldane model. In our earlier study (Chen, et al. (2017); Fig. 3), we show that the approximations can be very good in many (or most) settings.

      We would like to use the modern analogy of gas-engine cars vs. electric-motor ones. The Haldane model and the WF model are as fundamentally different concepts as the driving mechanisms of gas-powered vs electric cars.  The old model is now facing many problems and the fixes are often not possible.  Some fixes are so complicated that one starts thinking about simpler solutions. The reservations are that we have invested so much in the old models which might be wasted by the switch. However, we are suggesting the integration of the WF and Haldane models. In this sense, the WF model has had many contributions which the new model gratefully inherits. This is true with the legacy of gas-engine cars inherited by EVs.

      The editors also issue the instruction: while the modified model yields intriguing theoretical predictions, the simulations and empirical analyses are incomplete to support the authors' claims. 

      We are thankful to the editors and reviewers for the thoughtful comments and constructive criticisms. We also appreciate the publishing philosophy of eLife that allows exchanges, debates and improvements, which are the true spirits of science publishing.

      References for the provisional author responses

      Chen Y, Tong D, Wu CI. 2017. A New Formulation of Random Genetic Drift and Its Application to the Evolution of Cell Populations. Mol. Biol. Evol. 34:2057-2064.

      Hou M, Shi J, Gong Z, Wen H, Lan Y, Deng X, Fan Q, Li J, Jiang M, Tang X, et al. 2023. Intra- vs. Interhost Evolution of SARS-CoV-2 Driven by Uncorrelated Selection-The Evolution Thwarted. Mol. Biol. Evol. 40.

      Kimura M, Crow JF. 1963. The measurement of effective population number. Evolution:279-288.

      Ruan Y, Luo Z, Tang X, Li G, Wen H, He X, Lu X, Lu J, Wu CI. 2021. On the founder effect in COVID-19 outbreaks: how many infected travelers may have started them all? Natl. Sci. Rev. 8:nwaa246.

      Ruan Y, Wen H, He X, Wu CI. 2021. A theoretical exploration of the origin and early evolution of a pandemic. Sci Bull (Beijing) 66:1022-1029.

      Review comments

      eLife assessment 

      This study presents a useful modification of a standard model of genetic drift by incorporating variance in offspring numbers, claiming to address several paradoxes in molecular evolution.

      It is unfortunate that the study fails to engage prior literature that has extensively examined the impact of variance in offspring number, implying that some of the paradoxes presented might be resolved within existing frameworks.

      We do not believe that the paradoxes can be resolved.

      In addition, while the modified model yields intriguing theoretical predictions, the simulations and empirical analyses are incomplete to support the authors' claims. 

      Public Reviews: 

      Reviewer #1 (Public Review): 

      Summary: 

      The authors present a theoretical treatment of what they term the "Wright-Fisher-Haldane" model, a claimed modification of the standard model of genetic drift that accounts for variability in offspring number, and argue that it resolves a number of paradoxes in molecular evolution. Ultimately, I found this manuscript quite strange.

      The notion of effective population size as inversely related to the variance in offspring number is well known in the literature, and not exclusive to Haldane's branching process treatment. However, I found the authors' point about variance in offspring changing over the course of, e.g. exponential growth fairly interesting, and I'm not sure I'd seen that pointed out before.

      Nonetheless, I don't think the authors' modeling, simulations, or empirical data analysis are sufficient to justify their claims. 

      Weaknesses: 

      I have several outstanding issues. First of all, the authors really do not engage with the literature regarding different notions of an effective population. Most strikingly, the authors don't talk about Cannings models at all, which are a broad class of models with non-Poisson offspring distributions that nonetheless converge to the standard Wright-Fisher diffusion under many circumstances, and to "jumpy" diffusions/coalescents otherwise (see e.g. Mohle 1998, Sagitov (2003), Der et al (2011), etc.). Moreover, there is extensive literature on effective population sizes in populations whose sizes vary with time, such as Sano et al (2004) and Sjodin et al (2005).

      Of course in many cases here the discussion is under neutrality, but it seems like the authors really need to engage with this literature more. 

      The most interesting part of the manuscript, I think, is the discussion of the Density Dependent Haldane model (DDH). However, I feel like I did not fully understand some of the derivation presented in this section, which might be my own fault. For instance, I can't tell if Equation 5 is a result or an assumption - when I attempted a naive derivation of Equation 5, I obtained E(K_t) = 1 + r/c*(c-n)*dt. It's unclear where the parameter z comes from, for example. Similarly, is equation 6 a derivation or an assumption? Finally, I'm not 100% sure how to interpret equation 7. I that a variance effective size at time t? Is it possible to obtain something like a coalescent Ne or an expected number of segregating sites or something from this? 

      Similarly, I don't understand their simulations. I expected that the authors would do individual-based simulations under a stochastic model of logistic growth, and show that you naturally get variance in offspring number that changes over time. But it seems that they simply used their equations 5 and 6 to fix those values. Moreover, I don't understand how they enforce population regulation in their simulations---is N_t random and determined by the (independent) draws from K_t for each individual? In that case, there's no "interaction" between individuals (except abstractly, since logistic growth arises from a model that assumes interactions between individuals). This seems problematic for their model, which is essentially motivated by the fact that early during logistic growth, there are basically no interactions, and later there are, which increases variance in reproduction. But their simulations assume no interactions throughout! 

      The authors also attempt to show that changing variance in reproductive success occurs naturally during exponential growth using a yeast experiment. However, the authors are not counting the offspring of individual yeast during growth (which I'm sure is quite hard). Instead, they use an equation that estimates the variance in offspring number based on the observed population size, as shown in the section "Estimation of V(K) and E(K) in yeast cells". This is fairly clever, however, I am not sure it is right, because the authors neglect covariance in offspring between individuals. My attempt at this derivation assumes that I_t | I_{t-1} = \sum_{I=1}^{I_{t-1}} K_{i,t-1} where K_{i,t-1} is the number of offspring of individual i at time t-1. Then, for example, E(V(I_t | I_{t-1})) = E(V(\sum_{i=1}^{I_{t-1}} K_{i,t-1})) = E(I_{t-1})V(K_{t-1}) + E(I_{k-1}(I_{k-1}-1))*Cov(K_{i,t-1},K_{j,t-1}). The authors have the first term, but not the second, and I'm not sure the second can be neglected (in fact, I believe it's the second term that's actually important, as early on during growth there is very little covariance because resources aren't constrained, but at carrying capacity, an individual having offspring means that another individuals has to have fewer offspring - this is the whole notion of exchangeability, also neglected in this manuscript). As such, I don't believe that their analysis of the empirical data supports their claim. 

      Thus, while I think there are some interesting ideas in this manuscript, I believe it has some fundamental issues:

      first, it fails to engage thoroughly with the literature on a very important topic that has been studied extensively. Second, I do not believe their simulations are appropriate to show what they want to show. And finally, I don't think their empirical analysis shows what they want to show. 

      References: 

      Möhle M. Robustness results for the coalescent. Journal of Applied Probability. 1998;35(2):438-447. doi:10.1239/jap/1032192859 

      Sagitov S. Convergence to the coalescent with simultaneous multiple mergers. Journal of Applied Probability. 2003;40(4):839-854. doi:10.1239/jap/1067436085 

      Der, Ricky, Charles L. Epstein, and Joshua B. Plotkin. "Generalized population models and the nature of genetic drift." Theoretical population biology 80.2 (2011): 80-99 

      Sano, Akinori, Akinobu Shimizu, and Masaru Iizuka. "Coalescent process with fluctuating population size and its effective size." Theoretical population biology 65.1 (2004): 39-48 

      Sjodin, P., et al. "On the meaning and existence of an effective population size." Genetics 169.2 (2005): 1061-1070 

      Reviewer #2 (Public Review): 

      Summary: 

      This theoretical paper examines genetic drift in scenarios deviating from the standard Wright-Fisher model. The authors discuss Haldane's branching process model, highlighting that the variance in reproductive success equates to genetic drift. By integrating the Wright-Fisher model with the Haldane model, the authors derive theoretical results that resolve paradoxes related to effective population size. 

      Strengths: 

      The most significant and compelling result from this paper is perhaps that the probability of fixing a new beneficial mutation is 2s/V(K). This is an intriguing and potentially generalizable discovery that could be applied to many different study systems. 

      The authors also made a lot of effort to connect theory with various real-world examples, such as genetic diversity in sex chromosomes and reproductive variance across different species. 

      Weaknesses: 

      One way to define effective population size is by the inverse of the coalescent rate. This is where the geometric mean of Ne comes from. If Ne is defined this way, many of the paradoxes mentioned seem to resolve naturally. If we take this approach, one could easily show that a large N population can still have a low coalescent rate depending on the reproduction model. However, the authors did not discuss Ne in light of the coalescent theory. This is surprising given that Eldon and Wakeley's 2006 paper is cited in the introduction, and the multiple mergers coalescent was introduced to explain the discrepancy between census size and effective population size, superspreaders, and reproduction variance - that said, there is no explicit discussion or introduction of the multiple mergers coalescent. 

      The Wright-Fisher model is often treated as a special case of the Cannings 1974 model, which incorporates the variance in reproductive success. This model should be discussed. It is unclear to me whether the results here have to be explained by the newly introduced WFH model, or could have been explained by the existing Cannings model. 

      The abstract makes it difficult to discern the main focus of the paper. It spends most of the space introducing "paradoxes". 

      The standard Wright-Fisher model makes several assumptions, including hermaphroditism, non-overlapping generations, random mating, and no selection. It will be more helpful to clarify which assumptions are being violated in each tested scenario, as V(K) is often not the only assumption being violated. For example, the logistic growth model assumes no cell death at the exponential growth phase, so it also violates the assumption about non-overlapping generations. 

      The theory and data regarding sex chromosomes do not align. The fact that \hat{alpha'} can be negative does not make sense. The authors claim that a negative \hat{alpha'} is equivalent to infinity, but why is that? It is also unclear how theta is defined. It seems to me that one should take the first principle approach e.g., define theta as pairwise genetic diversity, and start with deriving the expected pair-wise coalescence time under the MMC model, rather than starting with assuming theta = 4Neu. Overall, the theory in this section is not well supported by the data, and the explanation is insufficient. 

      {Alpha and alpha' can both be negative.  X^2 = 0.47 would yield x = -0.7}

      Reviewer #3 (Public Review): 

      Summary: 

      Ruan and colleagues consider a branching process model (in their terminology the "Haldane model") and the most basic Wright-Fisher model. They convincingly show that offspring distributions are usually non-Poissonian (as opposed to what's assumed in the Wright-Fisher model), and can depend on short-term ecological dynamics (e.g., variance in offspring number may be smaller during exponential growth). The authors discuss branching processes and the Wright-Fisher model in the context of 3 "paradoxes": (1) how Ne depends on N might depend on population dynamics; (2) how Ne is different on the X chromosome, the Y chromosome, and the autosomes, and these differences do match the expectations base on simple counts of the number of chromosomes in the populations; (3) how genetic drift interacts with selection. The authors provide some theoretical explanations for the role of variance in the offspring distribution in each of these three paradoxes. They also perform some experiments to directly measure the variance in offspring number, as well as perform some analyses of published data. 

      Strengths: 

      (1) The theoretical results are well-described and easy to follow. 

      (2) The analyses of different variances in offspring number (both experimentally and analyzing public data) are convincing that non-Poissonian offspring distributions are the norm. 

      (3) The point that this variance can change as the population size (or population dynamics) change is also very interesting and important to keep in mind. 

      (4) I enjoyed the Density-Dependent Haldane model. It was a nice example of the decoupling of census size and effective size. 

      Weaknesses: 

      (1) I am not convinced that these types of effects cannot just be absorbed into some time-varying Ne and still be well-modeled by the Wright-Fisher process. 

      (2) Along these lines, there is well-established literature showing that a broad class of processes (a large subset of Cannings' Exchangeable Models) converge to the Wright-Fisher diffusion, even those with non-Poissonian offspring distributions (e.g., Mohle and Sagitov 2001). E.g., equation (4) in Mohle and Sagitov 2001 shows that in such cases the "coalescent Ne" should be (N-1) / Var(K), essentially matching equation (3) in the present paper. 

      (3) Beyond this, I would imagine that branching processes with heavy-tailed offspring distributions could result in deviations that are not well captured by the authors' WFH model. In this case, the processes are known to converge (backward-in-time) to Lambda or Xi coalescents (e.g., Eldon and Wakely 2006 or again in Mohle and Sagitov 2001 and subsequent papers), which have well-defined forward-in-time processes. 

      (4) These results that Ne in the Wright-Fisher process might not be related to N in any straightforward (or even one-to-one) way are well-known (e.g., Neher and Hallatschek 2012; Spence, Kamm, and Song 2016; Matuszewski, Hildebrandt, Achaz, and Jensen 2018; Rice, Novembre, and Desai 2018; the work of Lounès Chikhi on how Ne can be affected by population structure; etc...) 

      (5) I was also missing some discussion of the relationship between the branching process and the Wright-Fisher model (or more generally Cannings' Exchangeable Models) when conditioning on the total population size. In particular, if the offspring distribution is Poisson, then conditioned on the total population size, the branching process is identical to the Wright-Fisher model. 

      (6) In the discussion, it is claimed that the last glacial maximum could have caused the bottleneck observed in human populations currently residing outside of Africa. Compelling evidence has been amassed that this bottleneck is due to serial founder events associated with the out-of-Africa migration (see e.g., Henn, Cavalli-Sforza, and Feldman 2012 for an older review - subsequent work has only strengthened this view). For me, a more compelling example of changes in carrying capacity would be the advent of agriculture ~11kya and other more recent technological advances. 

      Recommendations for the authors: 

      Reviewing Editor Comments: 

      The reviewers recognize the value of this model and some of the findings, particularly results from the density-dependent Haldane model. However, they expressed considerable concerns with the model and overall framing of this manuscript.

      First, all reviewers pointed out that the manuscript does not sufficiently engage with the extensive literature on various models of effective population size and genetic drift, notably lacking discussion on Cannings models and related works.

      Second, there is a disproportionate discussion on the paradoxes, yet some of the paradoxes might already be resolved within current theoretical frameworks. All three reviewers found the modeling and simulation of the yeast growth experiment hard to follow or lacking justification for certain choices. The analysis approach of sex chromosomes is also questioned. 

      The reviewers recommend a more thorough review of relevant prior literature to better contextualize their findings. The authors need to clarify and/or modify their derivations and simulations of the yeast growth experiment to address the identified caveats and ensure robustness. Additionally, the empirical analysis of the sex chromosome should be revisited, considering alternative scenarios rather than relying solely on the MSE, which only provides a superficial solution. Furthermore, the manuscript's overall framing should be adjusted to emphasize the conclusions drawn from the WFH model, rather than focusing on the "unresolved paradoxes", as some of these may be more readily explained by existing frameworks. Please see the reviewers' overall assessment and specific comments. 

      Reviewer #2 (Recommendations For The Authors): 

      In the introduction -- "Genetic drift is simply V(K)" -- this is a very strong statement. You can say it is inversely proportional to V(K), but drift is often defined based on changes in allele frequency. 

      Page 3 line 86. "sexes is a sufficient explanation."--> "sex could be a sufficient explanation" 

      The strongest line of new results is about 2s/V(K). Perhaps, the paper could put more emphasis on this part and demonstrate the generality of this result with a different example. 

      The math notations in the supplement are not intuitive. e.g., using i_k and j_k as probabilities. I also recommend using E[X] and V[X]for expectation and variance rather than \italic{E(X)} to improve the readability of many equations. 

      Eq A6, A7, While I manage to follow, P_{10}(t) and P_{10} are not defined anywhere in the text. 

      Supplement page 7, the term "probability of fixation" is confusing in a branching model. 

      E.q. A 28. It is unclear eq. A.1 could be used here directly. Some justification would be nice. 

      Supplement page 17. "the biological meaning of negative..". There is no clear justification for this claim. As a reader, I don't have any intuition as to why that is the case.

    1. My fiance got a white Adler Tippa recently, but is unsure of the exact model or year. We looked up the serial number but nothing has come up even on the database. The Tippa plate just says Tippa, not Adler Tippa, so it can't be too old. Any ideas? Serial number: 10148440

      reply to u/DinoPup87 at https://new.reddit.com/r/typewriters/comments/1efzeor/adler_tippa_id/

      It's a common misconception that the database lists all serial numbers.

      You'll need to identify the make (and preferably the model) to search the database. Then you'll want to look at the serial numbers which your serial number appears between to be able to identify the year (or month if the data is granular enough) your machine was made. Reading the notes at the header of each page will give you details for how best to read and interpret the charts for each manufacturer. Notes and footnotes will provide you with additional details when available.

      You can then compare your machine against others which individuals have photographed and uploaded to the database. Feel free to add your typewriter as an example by making an account of your own. Doing this is sure to help researches and other enthusiasts in the future. Don't forget photos of your manual, tools which came with your machine, your case, and original dated purchase receipts if you have them.

    1. needs a new ink ribbon

      Chances are that you've got your original metal spools, and if so, definitely keep them. You can make a quick measurement, but I'm guessing you're going to want 1/2" or 13mm wide universal ribbon.

      You can buy this in many places and in various color combinations (if you have a bichrome machine—look for a black/white/red switch which can usually be found on the front of your machine) for just a few dollars for 16 yards or about 14 meters to fill up a 2 inch diameter spool. Often it will come on cheap universal plastic spools which you can use to wind onto your own original metal spools if necessary.

      Some machines often make use of proprietary mechanisms or geometry on their spools to effectuate the auto-reverse mechanism of the machine (though you'd have to check on your particular unit). Many machines after the 40s used small grommets on the ribbon itself to trigger the auto-reverse mechanism. If yours doesn't, you can trim these off with scissors as you spool the ribbon onto your machine if you're worried they'll get in the way.

      Some smaller ultra-portables can and often do use smaller diameter spools which only fit 12 yards of ribbon, but you can always cut your ribbon down from bigger spools if necessary.

      A few good sources of ribbon can be found at https://site.xavier.edu/polt/typewriters/tw-faq.html#q1.

      If you don't have the original spools and the cheap plastic universal ones don't work on yours, you can find replacements via https://www.ribbonsunlimited.com/ or by calling around to repair shops which may have extras https://site.xavier.edu/polt/typewriters/tw-manuals.html


      Incidentally, having your typewriter make and model as well as serial number can be helpful. You can often identify the model via https://typewriterdatabase.com/ if it's not on your typewriter directly. I'm guessing from the 2Y5852 that you've got a Good Companion No. 2 circa 1942, but you can track that down by looking at the database and individual galleries with photos.

      If you don't have one already, you might find a manual for your machine (or one very similar to it) at https://site.xavier.edu/polt/typewriters/tw-manuals.html

      reply to u/Fancy_Temporary_5902 at https://new.reddit.com/r/typewriters/comments/1eg176q/im_trying_to_id_a_typewriter_of_my_dads_as_it/

    1. Author response:

      The following is the authors’ response to the current reviews.

      (1) Though we cannot survey all mutants, our observation that 774 genetically diverse adaptive mutants converge at the level of phenotype is important. It adds to growing evidence (see PMID33263280, PMID37437111, PMID22282810, PMID25806684) that the genetic basis of adaptation is not as diverse as the phenotypic basis. This convergence could make evolution more predictable.

      (2) Previous fitness competitions using this specific barcode system have been run for greater than 25 generations (PMID33263280, PMID27594428, PMID37861305, PMID27594428). We measure fitness per cycle, rather than per generation, so our fitness advantages are comparable to those in the aforementioned studies, including Venkataram and Dunn et al. (PMID27594428).

      (3) Our results remain the same upon removing the ~150 lineages with the noisiest fitness inferences, including those the reviewer mentions (see Figure S7).

      (4) We agree that there are likely more than the 6 clusters that we validated with follow-up studies (see Discussion). The important point is that we see a great deal of convergence in the behavior of diverse adaptive mutants.

      (5) The growth curves requested by the reviewer were included in our original manuscript; several more were added in the revision (see Figures 5D, 5E, 7D, S11B, S11C).


      The following is the authors’ response to the original reviews.

      Public Reviews.

      Reviewer #1 (Public Review): 

      Summary: 

      In their manuscript, Schmidlin, Apodaca, et al try to answer fundamental questions about the evolution of new phenotypes and the trade-offs associated with this process. As a model, they use yeast resistance to two drugs, fluconazole and radicicol. They use barcoded libraries of isogenic yeasts to evolve thousands of strains in 12 different environments. They then measure the fitness of evolved strains in all environments and use these measurements to examine patterns in fitness trade-offs. They identify only six major clusters corresponding to different trade-off profiles, suggesting the vast genotypic landscape of evolved mutants translates to a highly constrained phenotypic space. They sequence over a hundred evolved strains and find that mutations in the same gene can result in different phenotypic profiles.  

      Overall, the authors deploy innovative methods to scale up experimental evolution experiments, and in many aspects of their approach tried to minimize experimental variation. 

      We thank the reviewer for this positive assessment of our work. We are happy that the reviewer noted what we feel is a unique strength of our approach: we scaled up experimental evolution by using DNA barcodes and by exploring 12 related selection pressures.  Despite this scaling up, we still see phenotypic convergence among the 744 adaptive mutants we study. 

      Weaknesses: 

      (1) One of the objectives of the authors is to characterize the extent of phenotypic diversity in terms of resistance trade-offs between fluconazole and radicicol. To minimize noise in the measurement of relative fitness, the authors only included strains with at least 500 barcode counts across all time points in all 12 experimental conditions, resulting in a set of 774 lineages passing this threshold. This corresponds to a very small fraction of the starting set of ~21 000 lineages that were combined after experimental evolution for fitness measurements. 

      This is a misunderstanding that we clarified in this revision. Our starting set did not include 21,000 adaptive lineages. The total number of unique adaptive lineages in this starting set is much lower than 21,000 for two reasons. 

      First, ~21,000 represents the number of single colonies we isolated in total from our evolution experiments. Many of these isolates possess the same barcode, meaning they are duplicates. Second, and perhaps more importantly, most evolved lineages do not acquire adaptive mutations, meaning that many of the 21,000 isolates are genetically identical to their ancestor. In our revised manuscript, we explicitly stated that these 21,000 isolated lineages do not all represent unique, adaptive lineages. We changed the word “lineages” to “isolates” where relevant in Figure 2 and the accompanying legend. And we have added the following sentence to the figure 2 legend (line 212), “These ~21,000 isolates do not represent as many unique, adaptive lineages because many either have the same barcode or do not possess adaptive mutations.”

      More broadly speaking, several previous studies have demonstrated that diverse genetic mutations converge at the level of phenotype and have suggested that this convergence makes adaptation more predictable (PMID33263280, PMID37437111, PMID22282810, PMID25806684). Most of these studies survey fewer than 774 mutants. Further, our study captures mutants that are overlooked in previous studies, such as those that emerge across subtly different selection pressures (e.g., 4 𝜇g/ml vs. 8 𝜇g/ml flu) and those that are undetectable in evolutions lacking DNA barcodes. Thus, while our experimental design misses some mutants (see next comment), it captures many others. Thus, we feel that “our work – showing that 774 mutants fall into a much smaller number of groups” is important because it “contributes to growing literature suggesting that the phenotypic basis of adaptation is not as diverse as the genetic basis (lines 176 - 178).”

      As the authors briefly remark, this will bias their datasets for lineages with high fitness in all 12 environments, as all these strains must be fit enough to maintain a high abundance. 

      We now devote 19 lines of text to discussing this bias (on lines 160 - 162, 278-284, and in more detail on 758 - 767).

      We walk through an example of a class of mutants that our study misses. One lines 759 - 763, we say, “our study is underpowered to detect adaptive lineages that have low fitness in any of the 12 environments. This is bound to exclude large numbers of adaptive mutants. For example, previous work has shown some FLU resistant mutants have strong tradeoffs in RAD (Cowen and Lindquist 2005). Perhaps we are unable to detect these mutants because their barcodes are at too low a frequency in RAD environments, thus they are excluded from our collection of 774.”

      In our revised version, we added more text earlier in the manuscript that explicitly discusses this bias. Lines 278 – 283 now read, “The 774 lineages we focus on are biased towards those that are reproducibly adaptive in multiple environments we study. This is because lineages that have low fitness in a particular environment are rarely observed >500 times in that environment (Figure S4). By requiring lineages to have high-coverage fitness measurements in all 12 conditions, we may be excluding adaptive mutants that have severe tradeoffs in one or more environments, consequently blinding ourselves to mutants that act via unique underlying mechanisms.”

      Note that while we “miss” some classes of mutants, we “catch” other classes that may have been missed in previous studies of convergence. For example, we observe a unique class of FLU-resistant mutants that primarily emerged in evolution experiments that lack FLU (Figure 3). Thus, we think that the unique design of our study, surveying 12 environments, allows us to make a novel contribution to the study of phenotypic convergence.

      One of the main observations of the authors is phenotypic space is constrained to a few clusters of roughly similar relative fitness patterns, giving hope that such clusters could be enumerated and considered to design antimicrobial treatment strategies. However, by excluding all lineages that fit in only one or a few environments, they conceal much of the diversity that might exist in terms of trade-offs and set up an inclusion threshold that might present only a small fraction of phenotypic space with characteristics consistent with generalist resistance mechanisms or broadly increased fitness. This has important implications regarding the general conclusions of the authors regarding the evolution of trade-offs. 

      We agree and discussed exactly the reviewer’s point about our inclusion threshold in the 19 lines of text mentioned previously (lines 160 - 162, 278-284, and 758 - 767). To add to this discussion, and avoid the misunderstanding the reviewer mentions, we added the following strongly-worded sentence to the end of the paragraph on lines 749 – 767 in our revised manuscript: “This could complicate (or even make impossible) endeavors to design antimicrobial treatment strategies that thwart resistance”. 

      More generally speaking, we set up our study around Figure 1, which depicts a treatment strategy that works best if there exists but a single type of adaptive mutant. Despite our inclusion threshold, we find there are at least 6 types of mutants. This diminishes hopes of designing simple multidrug strategies like Figure 1. Our goal is to present a tempered and nuanced discussion of whether and how to move forward with designing multidrug strategies, given our observations. On one hand, we point out how the phenotypic convergence we observe is promising. But on the other hand, we also point out how there may be less convergence than meets the eye for various reasons including the inclusion threshold the reviewer mentions (lines 749 - 767).

      We have made several minor edits to the text with the goal of providing a more balanced discussion of both sides. For example, we added the words, “may yet” to the following sentences on lines 32 – 36 of the abstract: “These findings, on one hand, demonstrate the difficulty in relying on consistent or intuitive tradeoffs when designing multidrug treatments. On the other hand, by demonstrating that hundreds of adaptive mutations can be reduced to a few groups with characteristic tradeoffs, our findings may yet empower multidrug strategies that leverage tradeoffs to combat resistance.”

      (2) Most large-scale pooled competition assays using barcodes are usually stopped after ~25 to avoid noise due to the emergence of secondary mutations. 

      The rate at which new mutations enter a population is driven by various factors such as the mutation rate and population size, so choosing an arbitrary threshold like 25 generations is difficult. 

      We conducted our fitness competition following previous work using the Levy/Blundell yeast barcode system, in which the number of generations reported varies from 32 to 40 (PMID33263280, PMID27594428, PMID37861305, see PMID27594428 for detailed calculation of the fraction of lineages biased by secondary mutations in this system). 

      The authors measure fitness across ~40 generations, which is almost the same number of generations as in the evolution experiment. This raises the possibility of secondary mutations biasing abundance values, which would not have been detected by the whole genome sequencing as it was performed before the competition assay. 

      Previous work has demonstrated that in this evolution platform, most mutations occur during the transformation that introduces the DNA barcodes (Levy et al. 2015). In other words, these mutations are already present and do not accumulate during the 40 generations of evolution. Therefore, the observation that we collect a genetically diverse pool of adaptive mutants after 40 generations of evolution is not evidence that 40 generations is enough time for secondary mutations to bias abundance values.

      We have added the following sentence to the main text to highlight this issue (lines 247 - 249): “This happens because the barcoding process is slightly mutagenic, thus there is less need to wait for DNA replication errors to introduce mutations (Levy et al. 2015; Venkataram et al. 2016).

      We also elaborate on this in the method section entitled, “Performing barcoded fitness competition experiments,” where we added a full paragraph to clarify this issue (lines 972 - 980).

      (3) The approach used by the authors to identify and visualize clusters of phenotypes among lineages does not seem to consider the uncertainty in the measurement of their relative fitness. As can be seen from Figure S4, the inter-replicate difference in measured fitness can often be quite large. From these graphs, it is also possible to see that some of the fitness measurements do not correlate linearly (ex.: Med Flu, Hi Rad Low Flu), meaning that taking the average of both replicates might not be the best approach.  Because the clustering approach used does not seem to take this variability into account, it becomes difficult to evaluate the strength of the clustering, especially because the UMAP projection does not include any representation of uncertainty around the position of lineages. This might paint a misleading picture where clusters appear well separate and well defined but are in fact much fuzzier, which would impact the conclusion that the phenotypic space is constricted. 

      Our noisiest fitness measurements correspond to barcodes that are the least abundant and thus suffer the most from stochastic sampling noise. These are also the barcodes that introduce the nonlinearity the reviewer mentions. We removed these from our dataset by increasing our coverage threshold from 500 reads to 5,000 reads. The clusters did not collapse, which suggests that they were not capturing this noise (Figure S7B).

      More importantly, we devoted 4 figures and 200 lines of text to demonstrating that the clusters we identified capture biologically meaningful differences between mutants (and not noise). We have modified the main text to point readers to figures 5 through 8 earlier, such that it is more apparent that the clustering analysis is just the first piece of our data demonstrating convergence at the level of phenotype.

      (4) The authors make the decision to use UMAP and a gaussian mixed model to cluster and represent the different fitness landscapes of their lineages of interest. Their approach has many caveats. First, compared to PCA, the axis does not provide any information about the actual dissimilarities between clusters. Using PCA would have allowed a better understanding of the amount of variance explained by components that separate clusters, as well as more interpretable components. 

      The components derived from PCA are often not interpretable. It’s not obvious that each one, or even the first one, will represent an intuitive phenotype, like resistance to fluconazole.  Moreover, we see many non-linearities in our data. For example, fitness in a double drug environment is not predicted by adding up fitness in the relevant single drug environments. Also, there are mutants that have high fitness when fluconazole is absent or abundant, but low fitness when mild concentrations are present. These types of nonlinearities can make the axes in PCA very difficult to interpret, plus these nonlinearities can be missed by PCA, thus we prefer other clustering methods. 

      Still, we agree that confirming our clusters are robust to different clustering methods is helpful. We have included PCA in the revised manuscript, plotting PC1 vs PC2 as Figure S9 with points colored according to the cluster assignment in figure 4 (i.e. using a gaussian mixture model). It appears the clusters are largely preserved.

      Second, the advantages of dimensional reduction are not clear. In the competition experiment, 11/12 conditions (all but the no drug, no DMSO conditions) can be mapped to only three dimensions: concentration of fluconazole, concentration of radicicol, and relative fitness. Each lineage would have its own fitness landscape as defined by the plane formed by relative fitness values in this space, which can then be examined and compared between lineages. 

      We worry that the idea stems from apriori notions of what the important dimensions should be. The biology of our system is unfortunately not intuitive. For example, it seems like this idea would miss important nonlinearities such as our observation that low fluconazole behaves more like a novel selection pressure than a dialed down version of high fluconazole. 

      Third, the choice of 7 clusters as the cutoff for the multiple Gaussian model is not well explained. Based on Figure S6A, BIC starts leveling off at 6 clusters, not 7, and going to 8 clusters would provide the same reduction as going from 6 to 7. This choice also appears arbitrary in Figure S6B, where BIC levels off at 9 clusters when only highly abundant lineages are considered. 

      We agree. We did not rely on the results of BIC alone to make final decisions about how many clusters to include. Another factor we considered were follow-up genotyping and phenotyping studies that confirm biologically meaningful differences between the mutants in each cluster (Figures 5 – 8). We now state this explicitly. Here is the modified paragraph where we describe how we chose a model with 7 clusters, from lines 436 – 446 of the revised manuscript:

      “Beyond the obvious divide between the top and bottom clusters of mutants on the UMAP, we used a gaussian mixture model (GMM) (Fraley and Raftery, 2003) to identify clusters. A common problem in this type of analysis is the risk of dividing the data into clusters based on variation that represents measurement noise rather than reproducible differences between mutants (Mirkin, 2011; Zhao et al., 2008). One way we avoided this was by using a GMM quality control metric (BIC score) to establish how splitting out additional clusters affected model performance (Figure S6). Another factor we considered were follow-up genotyping and phenotyping studies that demonstrate biologically meaningful differences between mutants in different clusters (Figures 5 – 8). Using this information, we identified seven clusters of distinct mutants, including one pertaining to the control strains, and six others pertaining to presumed different classes of adaptive mutant (Figure 4D). It is possible that there exist additional clusters, beyond those we are able to tease apart in this study.”

      This directly contradicts the statement in the main text that clusters are robust to noise, as more a stringent inclusion threshold appears to increase and not decrease the optimal number of clusters. Additional criteria to BIC could have been used to help choose the optimal number of clusters or even if mixed Gaussian modeling is appropriate for this dataset. 

      We are under the following impression: If our clustering method was overfitting, i.e. capturing noise, the optimal number of clusters should decrease when we eliminate noise. It increased. In other words, the observation that our clusters did not collapse (i.e.

      merge) when we removed noise suggests these clusters were not capturing noise. 

      Most importantly, our validation experiments, described below, provide additional evidence that our clusters capture meaningful differences between mutants (and not noise).  

      (5) Large-scale barcode sequencing assays can often be noisy and are generally validated using growth curves or competition assays. 

      Some types of bar-seq methods, in particular those that look at fold change across two time points, are noisier than others that look at how frequency changes across multiple timepoints (PMID30391162). Here, we use the less noisy method. We also reduce noise by using a stricter coverage threshold than previous work (e.g., PMID33263280), and by excluding batch effects by performing all experiments simultaneously, since we found this to be effective in our previous work (PMID37237236). 

      Perhaps also relevant is that the main assay we use to measure fitness has been previously validated (PMID27594428) and no subsequent study using this assay validates using the methods suggested above (see PMID37861305, PMID33263280, PMID31611676, PMID29429618, PMID37192196, PMID34465770, PMID33493203). Similarly, bar-seq has been used, without the suggested validation, to demonstrate that the way some mutant’s fitness changes across environments is different from other mutants (PMID33263280, PMID37861305, PMID31611676, PMID33493203, PMID34596043). This is the same thing that we use bar-seq to demonstrate. 

      For all of these reasons above, we are hesitant to confirm bar-seq itself as a valid way to infer fitness. It seems this is already accepted as a standard in our field. However, please see below.

      Having these types of results would help support the accuracy of the main assay in the manuscript and thus better support the claims of the authors. 

      While we don’t agree that fitness measurements obtained from this bar-seq assay generally require validation, we do agree that it is important to validate whether the mutants in each of our 6 clusters indeed are different from one another in meaningful ways.

      Our manuscript has 4 figures (5 - 8) and over 200 lines of text dedicated to validating whether our clusters capture reproducible and biologically meaningful differences between mutants. In the revised manuscript, we added additional validation experiments, such that three figures (Figures 5, 7 and S11) now involve growth curves, as the reviewer requested. 

      Below, we walk through the different types of validation experiments that are present in our manuscript, including those that were added in this revision.

      (1) Mutants from different clusters have different growth curves: In our original manuscript, we measured growth curves corresponding to a fitness tradeoff that we thought was surprising. Mutants in clusters 4 and 5 both have fitness advantages in single drug conditions. While mutants from cluster 4 also are advantageous in the relevant double drug conditions, mutants from cluster 5 are not! We validated these different behaviors by studying growth curves for a mutant from each cluster (Figures 7 and S11), finding that mutants from different clusters have different growth curves. In the revised manuscript, we added growth curves for 6 additional mutants (3 from cluster 1 and 3 from cluster 3), demonstrating that only the cluster 1 mutants have a tradeoff in high concentrations of fluconazole (see Figure 5D & 5E). In sum, this work demonstrates that mutants from different clusters have predictable differences in their growth phenotypes.

      (2) Mutants from different clusters have different evolutionary origins: In our original manuscript, we came up with a novel way to ask whether the clusters capture different types of adaptive mutants. We asked whether the mutants in each cluster originate from different evolution experiments. They often do (see pie charts in Figures 5, 6, 7, 8). In the revised manuscript, we extended this analysis to include mutants from cluster 1. Cluster 1 is defined by high fitness in low fluconazole that declines with increasing fluconazole. In our revised manuscript, we show that cluster 1 lineages were overwhelmingly sampled from evolutions conducted in our lowest concentration of fluconazole (see pie chart in new Figure 5A). No other cluster’s evolutionary history shows this pattern (compare to pie charts in figures 6, 7, and 8).

      **These pie charts also provide independent confirmation supporting the fitness tradeoffs observed for each cluster in figure 4E. For example, mutants in cluster 5 appear to have a tradeoff in a particular double drug condition (HRLF), and the pie charts confirm that they rarely originate from that evolution condition. This differs from cluster 4 mutants, which do not have a fitness tradeoff in HRLF, and are more likely to originate from that environment (see purple pie slice in figure 7). Additional cases where results of evolution experiments (pie charts) confirm observed fitness tradeoffs are discussed in the manuscript on lines 320 – 326, 594 – 598, 681 – 685.

      (3) Mutants from each cluster often fall into different genes: We sequenced many of these mutants and show that mutants in the same gene are often found in the same cluster. For example, all 3 IRA1 mutants are in cluster 6 (Fig 8), both GPB2 mutants are in cluster 4 (Figs 7 & 8), and 35/36 PDR mutants are in either cluster 2 or 3 (Figs 5 & 6). 

      (4) Mutants from each cluster have behaviors previously observed in the literature: We compared our sequencing results to the literature and found congruence. For example, PDR mutants are known to provide a fitness benefit in fluconazole and are found in clusters that have high fitness in fluconazole (lines 485 - 491). Previous work suggests that some mutations to PDR have different tradeoffs than others, which corresponds to our finding that PDR mutants fall into two separate clusters (lines 610 - 612). IRA1 mutants were previously observed to have high fitness in our “no drug” condition and are found in the cluster that has the highest fitness in the “no drug” condition (lines 691 - 696). Previous work even confirms the unusual fitness tradeoff we observe where IRA1 and other cluster 6 mutants have low fitness only in low concentrations of fluconazole (lines 702 - 704).

      (5) Mutants largely remain in their clusters when we use alternate clustering methods:  In our original manuscript, we performed various different re-clustering and/or normalization approaches on our data (Fig 6, S5, S7, S8, S10). The clusters of mutants that we observe in figure 4 do not change substantially when we re-cluster the data. In our revised manuscript, we added another clustering method: principal component analysis (PCA) (Fig S9).  Again, we found that our clusters are largely preserved.

      While these experiments demonstrate meaningful differences between the mutants in each cluster, important questions remain. For example, a long-standing question in biology centers on the extent to which every mutation has unique phenotypic effects versus the extent to which scientists can predict the effects of some mutations from other similar mutations. Additional studies on the clusters of mutants discovered here will be useful in deepening our understanding of this topic and more generally of the degree of pleiotropy in the genotype-phenotype map.

      Reviewer #2 (Public Review): 

      Summary: 

      Schmidlin & Apodaca et al. aim to distinguish mutants that resist drugs via different mechanisms by examining fitness tradeoffs across hundreds of fluconazole-resistant yeast strains. They barcoded a collection of fluconazole-resistant isolates and evolved them in different environments with a view to having relevance for evolutionary theory, medicine, and genotypephenotype mapping. 

      Strengths: 

      There are multiple strengths to this paper, the first of which is pointing out how much work has gone into it; the quality of the experiments (the thought process, the data, the figures) is excellent. Here, the authors seek to induce mutations in multiple environments, which is a really large-scale task. I particularly like the attention paid to isolates with are resistant to low concentrations of FLU. So often these are overlooked in favour of those conferring MIC values >64/128 etc. What was seen is different genotype and fitness profiles. I think there's a wealth of information here that will actually be of interest to more than just the fields mentioned (evolutionary medicine/theory). 

      We are grateful for this positive review. This was indeed a lot of work! We are happy that the reviewer noted what we feel is a unique strength of our manuscript: that we survey adaptive isolates across multiple environments, including low drug concentrations.  

      Weaknesses: 

      Not picking up low fitness lineages - which the authors discuss and provide a rationale as to why. I can completely see how this has occurred during this research, and whilst it is a shame I do not think this takes away from the findings of this paper. Maybe in the next one! 

      We thank the reviewer for these words of encouragement and will work towards catching more low fitness lineages in our next project.

      In the abstract the authors focus on 'tradeoffs' yet in the discussion they say the purpose of the study is to see how many different mechanisms of FLU resistance may exist (lines 679-680), followed up by "We distinguish mutants that likely act via different mechanisms by identifying those with different fitness tradeoffs across 12 environments". Whilst I do see their point, and this is entirely feasible, I would like a bit more explanation around this (perhaps in the intro) to help lay-readers make this jump. The remainder of my comments on 'weaknesses' are relatively fixable, I think: 

      We have expanded the introduction, in particular lines 129 – 157 of the revised manuscript, to walk readers through the connection between fitness tradeoffs and molecular mechanisms. For example, here is one relevant section of new text from lines 131 - 136: “The intuition here is as follows. If two groups of drug resistant mutants have different fitness tradeoffs, it could mean that they provide resistance through different underlying mechanisms. Alternatively, both could provide drug resistance via the same mechanism, but some mutations might also affect fitness via additional mechanisms (i.e. they might have unique “side-effects” at the molecular level) resulting in unique fitness tradeoffs in some environments.”

      In the introduction I struggle to see how this body of research fits in with the current literature, as the literature cited is a hodge-podge of bacterial and fungal evolution studies, which are very different! So example, the authors state "previous work suggests that mutants with different fitness tradeoffs may affect fitness through different molecular mechanisms" (lines 129-131) and then cite three papers, only one of which is a fungal research output. However, the next sentence focuses solely on literature from fungal research. Citing bacterial work as a foundation is fine, but as you're using yeast for this I think tailoring the introduction more to what is and isn't known in fungi would be more appropriate. It would also be great to then circle back around and mention monotherapy vs combination drug therapy for fungal infections as a rationale for this study. The study seems to be focused on FLU-resistant mutants, which is the first-line drug of choice, but many (yeast) infections have acquired resistance to this and combination therapy is the norm. 

      We ourselves are broadly interested in the structure of the genotype-phenotype-fitness map (PMID33263280, PMID32804946). For example, we are interested in whether diverse mutations converge at the level of phenotype and fitness. Figure 1A depicts a scenario with a lot of convergence in that all adaptive mutations have the same fitness tradeoffs.

      The reason we cite papers from yeast, as well as bacteria and cancer, is that we believe general conclusions about the structure of the genotype-phenotype-fitness map apply broadly. For example, the sentence the reviewer highlights, “previous work suggests that mutants with different fitness tradeoffs may affect fitness through different molecular mechanisms” is a general observation about the way genotype maps to fitness. So, we cited papers from across the tree of life to support this sentence.  And in the next sentence, where we cite 3 papers focusing solely on fungal research, we cite them because they are studies about the complexity of this map. Their conclusions, in theory, should also apply broadly, beyond yeast.

      On the other hand, because we study drug resistant mutations, we hope that our dataset and observations are of use to scientists studying the evolution of resistance. We use our introduction to explain how the structure of the genotype-phenotype-fitness map might influence whether a multidrug strategy is successful (Figure 1).

      We are hesitant to rework our introduction to focus more specifically on fungal infections as this is not our primary area of expertise.

      Methods: Line 769 - which yeast? I haven't even seen mention of which species is being used in this study; different yeast employ different mechanisms of adaptation for resistance, so could greatly impact the results seen. This could help with some background context if the species is mentioned (although I assume S. cerevisiae). 

      In the revised manuscript, we have edited several lines (line 95, 186, 822) to state the organism this work was done with is Saccharomyces cerevisiae. 

      In which case, should aneuploidy be considered as a mechanism? This is mentioned briefly on line 556, but with all the sequencing data acquired this could be checked quickly? 

      We like this idea and we are working on it, but it is not straightforward. The reviewer is correct in that we can use the sequencing data that we already have. But calling aneuploidy with certainty is tough because its signal can be masked by noise. In other words, some regions of the genome may be sequenced more than others by chance.

      Given this is not straightforward, at least not for us, this analysis will likely have to wait for a subsequent paper. 

      I think the authors could be bolder and try and link this to other (pathogenic) yeasts. What are the implications of this work on say, Candida infections? 

      Perhaps because our background lies in general study of the genotype-phenotype map, we are hesitant about making bold assertions about how our work might apply to pathogenic yeasts. We are hopeful that our work will serve as a stepping-stone such that scientists from that community can perhaps make (and test) such statements.   

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors): 

      I found the ideas and the questions asked in this manuscript to be interesting and ambitious. The setup of the evolution and fitness competition experiments was well poised to answer them, but the analysis of the data is not currently enough to properly support the claims made. I would suggest revising the analysis to address the weaknesses raised in the public review and if possible, adding some more experimental validations. As you already have genome sequencing data showing the causal mutation for many mutants across the different clusters, it should be possible for you to reconstruct some of the strains and test validate their phenotypes and cluster identity. 

      Yes, this is possible. We added more validation experiments (see figure 5). We already had quite a few validation experiments (figures 5 - 8 and lines 479 - 718), but we did not clearly highlight the significance of these analyses in our original manuscript. Therefore, we modified the text in our revised manuscript in various places to do so. For example, we now make clearer that we jointly use BIC scores as well as validation experiments to decide how many clusters to describe (lines 436 - 446). We also make clearer that our clustering analysis is only the first step towards identifying groups of mutants with similar tradeoffs by using words and phrases like, “we start by” (line 411) and “preliminarily” (line 448) when discussing the clustering analysis.  We also point readers to all the figures describing our validation experiments earlier (line 443), and list these experiments out in the discussion (lines 738 - 741).

      Also, please deposit your genome sequencing data in a public database (I am not sure I saw it mentioned anywhere). 

      We have updated line 1088 of the methods section to include this sentence: “Whole genome sequences were deposited in GenBank under SRA reference PRJNA1023288.”

      Reviewer #2 (Recommendations For The Authors):

      I don't think the figures or experiments can be improved upon, they are excellent. There are a few times I feel things are written in a rather confusing way and could be explained better, but also I feel there are places the authors jump from one thing to another really quickly and the reader (who might not be an expert in this area) will struggle to keep up. For example: 

      Explaining what RAD is - it is introduced in the methods, but what it is, is not really explained. 

      Since the introduction is already very long, we chose not to explain radicicol’s mechanism of action here. Instead, we bring this up later on lines 614 – 621 when it becomes relevant.

      More generally, in response to this advice and that from reviewer 1, we also added text to various places in the manuscript to help explain our work more clearly. In particular, we clarified the significance of our validation experiments and various important methodological details (see above). We also better explained the connection between fitness tradeoffs and mechanisms (see above) and added more details about the potential use cases of our approach (lines 142 – 150).

      The abstract states "some of the groupings we find are surprising. For example, we find some mutants that resist single drugs do not resist their combination, and some mutants to the same gene have different tradeoffs than others". Firstly, this sentence is a bit confusing to read but if I've read it as intended, then is it really surprising? It's difficult for organisms (bacteria and fungi) to develop multiple beneficial mutations conferring drug resistance on the same background, hence why combination antifungal drug therapy is often used to treat infections. 

      This is a place where brevity got in the way of clarity. We added a bit of text to make clear why we were surprised. Specifically, we were surprised because not all mutants behave the same. Some resist single drugs AND their combination. Some resist single drugs but not their combination. The sentence in the abstract now reads, “For example, we find some mutants that resist single drugs do not resist their combination, while others do. And some mutants to the same gene have different tradeoffs than others.”

    1. There is a disturbing new trend happening in Latin America, specifically in Colombia, and that is:

      0:07 Men being attracted by extremely beautiful, sexy Colombian and Latin American women and then being drugged, robbed, killed, overdosed, kidnapped, all the crimes that you can possibly think of.

      0:21 And a lot of men come to Latin America and thinking that, well, women are just gonna throw themselves at me. And to a certain extent, it is easier than dating in the United States and Germany and all these western countries, and it is better. Women are more beautiful, women are more feminine. But you have to be extremely careful. You have to know where you are.

      0:40 Recently, I've seen story after story after story, and this is an alarming trend because these criminals are getting harder. They're getting harsher on their crimes, they're getting more sadistic, and they're also planning a lot better.

      0:53 One recent case that I heard of was a German guy who went to Colombia, and on his second day, one girl that he met invited him to cook some food at home, cook some delicious Colombian food, and he said, why not? I'm probably not gonna bang this girl, but I'm gonna go home to her so she can cook me some food. She slipped scopolamine. a drug. into a drink, gave it to him, and then a few hours later, he woke up with his money gone, everything stolen, and even his crypto was stolen.

      1:28 They figured out how to get access to his crypto. They stole $15,000, which if you're watching this, you're a wealthy individual, you might think, well, 15K, whatever. But if you have 15 million in your crypto account or your Binance account or in your bank account, they will figure out how to steal it from you.

      1:42 Scopolamine is a drug that basically makes you into a little slave, into a little servant, and you'll do whatever the attacker wants. They tell you, go tell the security and tell them that I'm your friend. That literally happens in Colombia.

      1:57 People get drugged, and then they go to their Airbnb, to their hotel, and they tell the security, that's my friend. Let them come with me. And they come with you, and they steal absolutely everything because it's called the devil's breath.

      2:08 It's a drug that essentially turns you into a zombie, and it has bans all over Latin America, and a lot of police forces in Latin America are trying to ban this drug. They're trying to control the amount of it that is produced, imported, and they have very strict penalties. If you traffic it, if you sell it, you go to prison for a long time.

      2:26 But in these countries, these women, they figured out how to get as much money as possible from these expats, from these tourists. You might meet them on Tinder, you might meet an absolutely beautiful girl, and she invites you to some coffee shop. And then after the coffee shop, you think, yeah, this is going so well. She says, let's go cook some food. Let's go meet some of my friends.

      2:46 Or in one case that I saw, she invited him with her private driver. She said, oh, I have a private driver. He can take us really nice places. It's not safe here. And the private driver was the guy's kidnapper, ended up kidnapping the guy.

      3:00 And there was also another story of a very famous Minnesota man, Asian man, but American, Asian-American. He went to Colombia. He met an absolutely beautiful girl. He showed her off to his family. This Asian American was then kidnapped by this girl, but not immediately, not on the first date. Multiple dates later, after he came back for a second time to Colombia. He met her first, went back home, told everybody about his beautiful girlfriend, came back to Colombia, and then he was kidnapped. The kidnappers asked for $2,000, which is ridiculous, and then they killed him anyway.

      3:42 And this can happen to you in Mexico, in Colombia, Brazil. You have to know where you're going. If you're in Mexico, don't go out past eight, 9:00 PM in bad areas. If you're coming to a place like Tulum, stay in the absolute safest areas. Don't think, ah, I know what I'm doing. Ah, they're exaggerating the crime.

      4:00 And especially if you're come to Latin America for dating or if you're using dating apps, if it's too good to be true, it literally is.

      4:09 If you meet a girl on Tinder and she has bikini pictures all over her profile, if you see that she's pushing hard to meet you, if you see that she's pushing hard to either go to your place or to go to her place. If she wants you to go to her place, it's a red flag anywhere in the world. Even in Russia.

      4:24 I heard of a story from a friend of mine. The guy was invited by the girl to her house, and he thought, oh, I'm getting laid with an extremely hot rushing girl. He went to her place. He got his kidney removed. He woke up the next morning, whoa, disoriented without his kidney.

      4:38 Well, of course, a girl isn't going to invite you to your place in a random country, especially if she doesn't speak your language. Most girls are not that slutty. That's probably not gonna happen to you, unless you're some footballer or a famous person.

      4:51 So you have to keep an eye out for red flags, and you always have to keep in mind that people will try to take advantage of you, especially if you don't speak the language.

      5:00 I speak native Spanish. I don't wear my Rolex. I don't wear expensive clothes when I'm going out in Mexico, in Colombia and Argentina. You just don't show off. You speak the language or you try to. If you're going to Brazil, learn Portuguese, because you are a target, especially if you're tall, white.

      5:20 I've seen many tall Americans, white as hell, white as paper, and they walk through Mexico like it's their front yard. You are a target. Don't wear expensive jewelry. Here in Tulum, Rolexes get stolen all the time. I was reading through many articles of gun robberies and overall armed violence, and they were all because a person had an extremely expensive something, either a camera, and I'm filming this in an area where they're actually building new buildings. There's security all over the place. It's called Selvazama, absolutely beautiful. They're gonna, well, they're gonna chop up all these trees and they're gonna build new buildings. It's quite safe here. It's actually the best. One of the safest place in Mexico, I would say.

      5:56 But if you're going through a rough area, especially if it's at night, you don't wanna have a Rolex. You don't want to have Gucci shoes. I'm wearing my New balance, my little shorts from Zara.

      6:08 You have to know where you are. It's not Dubai. It's not Miami. You have to always be aware, and if it's too good to be true, it probably is when dating these Latino women.

      6:18 Now, my full game plan. I was born and raised in Puerto Rico, one of the least safest places in the world. I've spent a lot of time in Colombia, in Mexico, in Dominican Republic, many places where people get robbed, they get stabbed, they get lost, they get kidnapped.

      6:31 My game plan, one, as I said, do not show off in any way. You think, oh, a Rolex is gonna get me laid. No, it's gonna get you kidnapped. Do not show off. Do not wear expensive clothes. Wear Zara, H&M, even if you're a billionaire, just wear the cheapest clothes possible while looking nice. You don't wanna look like a homeless person, but if you're going through a rough area, it's better to look like a homeless person, so that the attacker thinks: This person is more poor than I am. Why am I gonna attack them?

      6:53 Second of all, if you have an expensive phone, like an iPhone, try to not use it that much outside, to be honest, or get a copy. For example, you could buy a second phone, like an iPhone 10 or an iPhone 11, and then you have your iPhone 15 at home. You use the other iPhone for just going outside, Google Maps, WhatsApp, get a local WhatsApp number.

      7:10 Again, speak the language. If you speak the language with a Colombian girl, she thinks twice about kidnapping you, about taking you somewhere. The taxi drivers will think twice about robbing you or putting a gun to your face because you speak the language. You know yourself around.

      7:25 And also make friends in these places. If you go to Mexico, know local people so that you can always keep them updated if everything is okay.

      7:32 If you're going to a place like I went to, Tijuana, very dangerous city in Mexico, next to the border, or in the border or at the border, with the United States, you wanna have people that you let them know every few hours how you are, or every day how you are.

      7:46 Hey, I'm doing great. I'm here in Tijuana. I stayed the night. Everything is fine, everything is fine. You just let them know that everything is fine. This is how Latino people keep themselves updated to see if everything is fine. My family's like that, hey, are you alive? Everything fine?

      7:59 We're not in a war zone, but this is how Latino culture is, because we know that shit happens. We know that people get kidnapped. We know that people get stabbed, so you wanna get adjusted to that culture.

      8:07 And overall, know the different areas of the city. Stay in the absolute best, safest possible area every time you go somewhere. Don't try to save a buck. Don't try to stay in the area with the most people with the highest chance of getting laid. And don't do anything stupid.

      8:21 Don't go to some cabaret. I see it on Reddit all the time, on the Mexico groups and on the the different Latin American groups, that people go to cabarets, erotic massage centers. They go to different nightclubs that they shouldn't be going to, and they get stabbed. They get robbed, they get kidnapped. You want to be vigilant.

      8:35 It's a great place. Latin America is beautiful, it's free. There's investment opportunities. There's land to buy, there's low taxes, there's beautiful women. It's one of the best areas of the world, but you have to know where you are.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      The study of human intelligence has been the focus of cognitive neuroscience research, and finding some objective behavioral or neural indicators of intelligence has been an ongoing problem for scientists for many years. Melnick et al, 2013 found for the first time that the phenomenon of spatial suppression in motion perception predicts an individual's IQ score. This is because IQ is likely associated with the ability to suppress irrelevant information. In this study, a high-resolution MRS approach was used to test this theory. In this paper, the phenomenon of spatial suppression in motion perception was found to be correlated with the visuo-spatial subtest of gF, while both variables were also correlated with the GABA concentration of MT+ in the human brain. In addition, there was no significant relationship with the excitatory transmitter Glu. At the same time, SI was also associated with MT+ and several frontal cortex FCs.

      Strengths:

      (1) 7T high-resolution MRS is used.

      (2) This study combines the behavioral tests, MRS, and fMRI.

      Weaknesses:

      (1) In the intro, it seems to me that the multiple-demand (MD) regions are the key in this study. However, I didn't see any results associated with the MD regions. Did I miss something?

      Thank you to the reviewer for pointing this out. After careful consideration, we agree with your point of view. According to the results of Melnick 2013, the motion surround suppression (SI) and the time thresholds of small and large gratings representing hMT+ functionality are correlated with Verbal Comprehension, Perceptual Reasoning, Working Memory, and Processing Speed Indicators, with correlation coefficients of 0.69, 0.47, 0.49, and 0.50, respectively. This suggests that hMT+ does have the potential to become the core of MD system. However, due to our results only delving into “the GABA-ergic inhibition in human MT predicts visuo-spatial intelligence mediated through the frontal cortex”, it is not yet sufficient to prove that hMT+is the core node of the MD system, we have adjusted the explanatory logic of the article. Briefly, we emphasize the de-redundancy of hMT+ in visual-spatial intelligence and the improvement of information processing efficiency, while weaken the significance of hMT+ in MD systems.

      (2) How was the sample size determined? Is it sufficient?

      Thank you to reviewer for pointing this out. We use G*power to determine our sample size. In the study by Melnick (2013), they reported a medium effect between SI and Perception Reasoning sub-ability (r=0.47). Here we use this r value as the correlation coefficient (ρ H1), setting the power at the commonly used threshold of 0.8 and the alpha error probability at 0.05. The required sample size is calculated to be 26. This ensures that our study has reasonable power to yield valid statistical results. Furthermore, compared to earlier within-subject studies like Schallmo et al.'s 2018 research, which used 22 datasets to examine GABA levels in MT+ and the early visual cortex (EVC), our study includes an enough dataset.

      (3) In Schallmo elife 2018, there was no correlation between GABA concentration and SI. How can we justify the different results different here?

      Thank reviewer for pointing this out. There are several differences between us:

      a. While the earlier study by Schallmo et al. (2018) employed 3T MRS, we utilize 7T MRS, enhancing our ability to detect and measure GABA with greater accuracy.

      b. Schallmo elife 2018 choose to use the bilateral hMT+ as the MRS measurement region while we use the left hMT+. The reason why we focus on left hMT+ are describe in reviewer 1. (6). Briefly, use of left MT/V5 as a target was motivated by studies demonstrating that left MT/V5 TMS is more effective at causing perceptual effects (Tadin et al., 2011).

      c. The resolution of MRS sequence in Schallmo elife 2018 is 3 cm isotropic voxel, while we apply 2 cm isotropic voxel. This helps us more precisely locate hMT+ and exclude more white matter signal.

      (4) Basically this study contains the data of SI, BDT, GABA in MT+ and V1, Glu in MT+ and V1-all 6 measurements. There should be 6x5/2 = 15 pairwise correlations. However, not all of these results are included in Figure 1 and supplementary 1-3. I understand that it is not necessary to include all figures. But I suggest reporting all values in one Table.

      We thank the reviewer for the good suggestion, we have made a correlation matrix to reporting all values in Figure Supplementary 9.

      (5) In Melnick (2013), the IQ scores were measured by the full set of WAIS-III, including all subtests. However, this study only used the visual spatial domain of gF. I wonder why only the visuo-spatial subtest was used not the full WAIS-III?

      We thank the reviewer for pointing this out. The decision was informed by Melnick’s findings which indicated high correlations between Surround suppression (SI) and the Verbal Comprehension, Perceptual Reasoning, Working Memory, and Processing Speed Indexes, with correlation coefficients of 0.69, 0.47, 0.49, and 0.50, respectively. It is well-established that the hMT+ region of the brain is a sensory cortex involved in visual perception processing (3D perception). Furthermore, motion surround suppression (SI), a specific function of hMT+, aligns closely with this region's activities. Given this context, the Perception Reasoning sub-ability was deemed to have the clearest mechanism for further exploration. Consequently, we selected the most representative subtest of Perception Reasoning—the Block Design Test—which primarily assesses 3D visual intelligence.

      (6) In the functional connectivity part, there is no explanation as to why only the left MT+ was set to the seed region. What is the problem with the right MT+?

      We thank the reviewer for pointing this out. The main reason is that our MRS ROI is the left hMT+, we would like to make different models’ ROI consistent to each other. Use of left MT/V5 as a target was motivated by studies demonstrating that left MT/V5 TMS is more effective at causing perceptual effects (Tadin et al., 2011).

      (7) In Melnick (2013), the authors also reported the correlation between IQ and absolute duration thresholds of small and large stimuli. Please include these analyses as well.

      We thank the reviewer for the good advice. Containing such result do help researchers compare the result between Melnick and us. We have made such figures in the revised version (Figure 3f, g).

      Reviewer #2 (Public Review):

      Summary:

      Recent studies have identified specific regions within the occipito-temporal cortex as part of a broader fronto-parietal, domain-general, or "multiple-demand" (MD) network that mediates fluid intelligence (gF). According to the abstract, the authors aim to explore the mechanistic roles of these occipito-temporal regions by examining GABA/glutamate concentrations. However, the introduction presents a different rationale: investigating whether area MT+ specifically, could be a core component of the MD network.

      Strengths:

      The authors provide evidence that GABA concentrations in MT+ and its functional connectivity with frontal areas significantly correlate with visuo-spatial intelligence performance. Additionally, serial mediation analysis suggests that inhibitory mechanisms in MT+ contribute to individual differences in a specific subtest of the Wechsler Adult Intelligence Scale, which assesses visuo-spatial aspects of gF.

      Weaknesses:

      (1) While the findings are compelling and the analyses robust, the study's rationale and interpretations need strengthening. For instance, Assem et al. (2020) have previously defined the core and extended MD networks, identifying the occipito-temporal regions as TE1m and TE1p, which are located more rostrally than MT+. Area MT+ might overlap with brain regions identified previously in Fedorenko et al., 2013, however the authors attribute these activations to attentional enhancement of visual representations in the more difficult conditions of their tasks. For the aforementioned reasons, It is unclear why the authors chose MT+ as their focus. A stronger rationale for this selection is necessary and how it fits with the core/extended MD networks.

      We really appreciate reviewer’s opinions. The reason why we focus on hMT+ is following: According to the results of Melnick 2013, the motion surround suppression (SI) and the time thresholds of small and large gratings representing hMT+ functionality are correlated with Verbal Comprehension, Perceptual Reasoning, Working Memory, and Processing Speed Indicators, with high correlation coefficients of 0.69, 0.47, 0.49, and 0.50, respectively. In addition, Fedorenko et al. 2013, the averaged MD activity region appears to overlap with hMT+. Based on these findings, we assume that hMT+ does have the potential to become the core of MD system.

      (2) Moreover, although the study links MT+ inhibitory mechanisms to a visuo-spatial component of gF, this evidence alone may not suffice to position MT+ as a new core of the MD network. The MD network's definition typically encompasses a range of cognitive domains, including working memory, mathematics, language, and relational reasoning. Therefore, the claim that MT+ represents a new core of MD needs to be supported by more comprehensive evidence.

      Thank reviewer for pointing this out. After careful consideration, we agree with your point of view. Due to our results only delving into visuo-spatial intelligence, it is not yet sufficient to prove that hMT is the core node of the MD system. We will adjust the explanatory logic of the article, that is, emphasizing the de-redundancy of hMT+in visual-spatial intelligence and the improvement of information processing efficiency, while weakening the significance of hMT+ in MD systems.

      Reviewer #3 (Public Review):

      Summary:

      This manuscript aims to understand the role of GABA-ergic inhibition in the human MT+ region in predicting visuo-spatial intelligence through a combination of behavioral measures, fMRI (for functional connectivity measurement), and MRS (for GABA/glutamate concentration measurement). While this is a commendable goal, it becomes apparent that the authors lack fundamental understanding of vision, intelligence, or the relevant literature. As a result, the execution of the research is less coherent, dampening the enthusiasm of the review.

      Strengths:

      (1) Comprehensive Approach: The study adopts a multi-level approach, i.e., neurochemical analysis of GABA levels, functional connectivity, and behavioral measures to provide a holistic understanding of the relationship between GABA-ergic inhibition and visuo-spatial intelligence.

      (2) Sophisticated Techniques: The use of ultra-high field magnetic resonance spectroscopy (MRS) technology for measuring GABA and glutamate concentrations in the MT+ region is a recent development.

      Weaknesses:

      Study Design and Hypothesis

      (1) The central hypothesis of the manuscript posits that "3D visuo-spatial intelligence (the performance of BDT) might be predicted by the inhibitory and/or excitation mechanisms in MT+ and the integrative functions connecting MT+ with the frontal cortex." However, several issues arise:

      (1.1) The Suppression Index depicted in Figure 1a, labeled as the "behavior circle," appears irrelevant to the central hypothesis.

      We thank the reviewer for pointing this out. In our study, the inhibitory mechanisms in hMT+ are conceptualized through two models: the neurotransmitter model and the behavioral model. The Suppression Index is essential for elucidating the local inhibitory mechanisms within the behavioral model. However, we acknowledge that our initial presentation in the introduction may not have clearly articulated our hypothesis, potentially leading to misunderstandings. We have revised the introduction to better clarify these connections and ensure the relevance of the Suppression Index is comprehensively understood.

      (1.2) The construct of 3D visuo-spatial intelligence, operationalized as the performance in the Block Design task, is inconsistently treated as another behavioral task throughout the manuscript, leading to confusion.

      We thank the reviewer for pointing this out. We acknowledge that our manuscript may have inconsistently presented this construct across different sections, causing confusion. To address this, we ensured a consistent description of 3D visuo-spatial intelligence in both the introduction and the discussion sections. But we maintained ‘Block Design task score' within the results section to help readers clarify which subtest we use.

      (1.3) The schematics in Figure 1a and Figure 6 appear too high-level to be falsifiable. It is suggested that the authors formulate specific and testable hypotheses and preregister them before data collection.

      We thank the reviewer for pointing this out. We have revised the Figure 1a and made it less abstract and more logical. For Figure 6, the schematic represents our theoretical framework of how hMT+ contributes to 3D visuo-spatial intelligence, we believe the elements within this framework are grounded in related theories and supported by evidence discussed in our results and discussions section, making them specific and testable.

      (2) Central to the hypothesis and design of the manuscript is a misinterpretation of a prior study by Melnick et al. (2013). While the original study identified a strong correlation between WAIS (IQ) and the Suppression Index (SI), the current manuscript erroneously asserts a specific relationship between the block design test (from WAIS) and SI. It should be noted that in the original paper, WAIS comprises Similarities, Vocabulary, Block design, and Matrix reasoning tests in Study 1, while the complete WAIS is used in Study 2. Did the authors conduct other WAIS subtests other than the block design task?

      Thank you for pointing this out. Reviewer #1 also asked this question, we copy the answers in here “The decision was informed by Melnick’s findings which indicated high correlations between Surround suppression (SI) and the Verbal Comprehension, Perceptual Reasoning, Working Memory, and Processing Speed Indexes, with correlation coefficients of 0.69, 0.47, 0.49, and 0.50, respectively. It is well-established that the hMT+ region of the brain is a sensory cortex involved in visual perception processing (3D perception). Furthermore, motion surround suppression (SI), a specific function of hMT+, aligns closely with this region's activities. Given this context, the Perception Reasoning sub-ability was deemed to have the clearest mechanism for further exploration. Consequently, we selected the most representative subtest of Perception Reasoning—the Block Design Test—which primarily assesses 3D visual intelligence.”

      (3) Additionally, there are numerous misleading references and unsubstantiated claims throughout the manuscript. As an example of misleading reference, "the human MT ... a key region in the multiple representations of sensory flows (including optic, tactile, and auditory flows) (Bedny et al., 2010; Ricciardi et al., 2007); this ideally suits it to be a new MD core." The two references in this sentence are claims about plasticity in the congenitally blind with sensory deprivation from birth, which is not really relevant to the proposal that hMT+ is a new MD core in healthy volunteers.

      Thank you for pointing this out. We have carefully read the corresponding references and considered the corresponding theories and agree with these comments. Due to our results only delving into “the GABA-ergic inhibition in human MT predicts visuo-spatial intelligence mediated by reverberation with frontal cortex”, it is not yet sufficient to prove that hMT+ is the core node of the MD system, we will adjust the explanatory logic of the article, that is, emphasizing the de redundancy of hMT+in visual-spatial intelligence and the improvement of information processing efficiency, while weakening the significance of hMT+ in MD systems. In addition, regarding the potential central role of hMT+ in the MD system, we agree with your view that research on hMT+ as a multisensory integration hub mainly focuses on developmental processes. Meanwhile, in adults, the MST region of hMT+ is considered a multisensory integration area for visual and vestibular inputs, which potentially supports the role of hMT+ in multitasking multisensory systems (Gu et al., J. Neurosci, 26(1), 73–85, 2006; Fetsch et al., Nat. Neurosci, 15, 146–154, 2012.). Further research could explore how other intelligence sub-ability such as working memory and language comprehension are facilitated by hMT+'s features.

      Another example of unsubstantiated claim: the rationale for selecting V1 as the control region is based on the assertion that "it mediates the 2D rather than 3D visual domain (Born & Bradley, 2005)". That's not the point made in the Born & Bradley (2005) paper on MT. It's crucial to note that V1 is where the initial binocular convergence occurs in cortex, i.e., inputs from both the right and left eyes to generate a perception of depth.

      Thank you for pointing this out. We acknowledge the inappropriate citation of "Born & Bradley, 2005," which focuses solely on the structure and function of the visual area MT. However, we believe that choosing hMT+ as the domain for 3D visual analysis and V1 as the control region is justified. Cumming and DeAngelis (Annu Rev Neurosci, 24:203–238.2001) state that binocular disparity provides the visual system with information about the three-dimensional layout of the environment, and the link between perception and neuronal activity is stronger in the extrastriate cortex (especially MT) than in the primary visual cortex. This supports our choice and emphasizes the relevance of hMT+ in our study. We have revised our reference in the revised version.

      Results & Discussion

      (1) The missing correlation between SI and BDT is crucial to the rest of the analysis. The authors should discuss whether they replicated the pattern of results from Melnick et al. (2013) despite using only one WAIS subtest.

      We thank for the reviewer’s suggestion. We have placed it in the main text (Figure 3e).

      (2) ROIs: can the authors clarify if the results are based on bilateral MT+/V1 or just those in the left hemisphere? Can the authors plot the MRS scan area in V1? I would be surprised if it's precise to V1 and doesn't spread to V2/3 (which is fine to report as early visual cortex).

      We thank for the reviewer’s suggestion. We have drawn the V1 ROI MRS scanning area (Figure supplement 1). Using the template, we checked the coverage of V1, V2, and V3. Although the MRS overlap regions extend to V2 (3%) and V3 (32%), the major coverage of the MRS scanning area is in V1, with 65% overlap across subjects.

      (3) Did the authors examine V1 FC with either the frontal regions and/or whole brain, as a control analysis? If not, can the author justify why V1 serves as the control region only in the MRS but not in FC (Figure 4) or the mediation analysis (Figure 5)? That seems a little odd given that control analyses are needed to establish the specificity of the claim to MT+

      We thank for the reviewer’s suggestion. We have done the V1 FC-behavior connection as control analysis (Figure supplement 7). Only positive correlations in the frontal area were detected, suggesting that in the 3D visuo-spatial intelligence task, V1 plays a role in feedforward information processing. However, hMT+, which showed specific negative correlations in the frontal, is involved in the inhibition mechanism. These results further emphasize the de-redundancy function of hMT+ in 3D visuo-spatial intelligence.

      Regarding the mediation analysis, since GABA/Glu concentration in V1 has no correlation with BDT score, it is not sufficient to apply mediation analysis.

      (4) It is not clear how to interpret the similarity or difference between panels a and b in Figure 4.

      We thank the reviewer for pointing this out. We have further interpreted the difference between a and b in the revised version. Panels a represents BDT score correlated hMT+-region FC, which is obviously involved in frontal cortex. While panels b represents SI correlated hMT+-region FC, which shows relatively less regions. The overlap region is what we are interested in and explain how local inhibitory mechanisms works in the 3D visuo-spatial intelligence. In addition, we have revised Figure 4 and point out the overlap region.

      (5) SI is not relevant to the authors‘ priori hypothesis, but is included in several mediation analyses. Can the authors do model comparisons between the ones in Figure 5c, d, and Figure S6? In other words, is SI necessary in the mediation model? There seem discrepancies between the necessity of SI in Figures 5c/S6 vs. Figure 5d.

      We thank the reviewer for highlighting this point. The relationship between the Suppression Index (SI) and our a priori hypotheses is elaborated in the response to reviewer 3, section (1). SI plays a crucial role in explicating how local inhibitory mechanisms, on the psychological level, function within the context of the 3D visuo-spatial task. Additionally, Figure 5c illustrates the interaction between the frontal cortex and hMT+, showing how the effects from the frontal cortex (BA46) on the Block Design Task are fully mediated by SI. This further underscores the significance of SI in our model.

      (6) The sudden appearance of "efficient information" in Figure 6, referring to the neural efficiency hypothesis, raises concerns. Efficient visual information processing occurs throughout the visual cortex, starting from V1. Thus, it appears somewhat selective to apply the neural efficiency hypothesis to MT+ in this context.

      We thank the reviewer for highlighting this point. There is no doubt that V1 involved in efficient visual information processing. However, in our result, the V1 GABA has no significant correlation between BDT score, suggesting that the V1 efficient processing might not sufficiently account for the individual differences in 3D visuo-spatial intelligence. Additionally, we will clarify our use of the neural efficiency hypothesis by incorporating it into the introduction of our paper to better frame our argument.

      Transparency Issues:

      (1) Don't think it's acceptable to make the claim that "All data needed to evaluate the conclusions in the paper are present in the paper and/or the Supplementary information". It is the results or visualizations of data analysis, rather than the raw data themselves, that are presented in the paper/supp info.

      We thank the reviewer for pointing this out. We realized that such expression would lead to confusion. We have deleted this expression.

      (2) No GitHub link has been provided in the manuscript to access the source data, which limits the reproducibility and transparency of the study.

      We thank the reviewer for pointing this out. We have attached the GitHub link in the revised version.

      Minor:

      "Locates" should be replaced with "located" throughout the paper. For example: "To investigate this issue, this study selects the human MT complex (hMT+), a region located at the occipito-temporal border, which represents multiple sensory flows, as the target brain area."

      We thank the reviewer for pointing this out. We have revised it.

      Use "hMT+" instead of "MT+" to be consistent with the term in the literature.

      We thank the reviewer for pointing this out. We agree to use hMT+ in the literature.

      "Green circle" in Figure 1 should be corrected to match its actual color.

      We thank the reviewer for pointing this out. We have revised it.

      The abbreviation for the Wechsler Adult Intelligence Scale should be "WAIS," not "WASI."

      We thank the reviewer for pointing this out. We have revised it.

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors):

      (1) The figures and tables should be substantially improved.

      We thank the reviewer for pointing this out. We have improved some of the figures’ quality.

      (2) Please explain the sample size, and the difference between Schallmo eLife 2018, and Melnick, 2013.

      We thank the reviewer for pointing this out. These questions are answered in the public review. We copy the answer in the public review.

      (2.1)  How was the sample size determined? Is it sufficient??

      Thank you to the reviewer for pointing this out. We use G*power to determine our sample size. In the study by Melnick (2013), they reported a medium effect between SI and Perception Reasoning sub-ability (r=0.47). Here we use this r value as the correlation coefficient (ρ H1), setting the power at the commonly used threshold of 0.8 and the alpha error probability at 0.05. The required sample size is calculated to be 26. This ensures that our study has adequate power to yield valid statistical results. Furthermore, compared to earlier within-subject studies like Schallmo et al.'s 2018 research, which used 22 subjects to examine GABA levels in MT+ and the early visual cortex (EVC), our study includes an enough dataset.

      (2.2)  In Schallmo elife 2018, there was no correlation between GABA concentration and SI. How can we justify the different results different here?

      Thank you to the reviewer for pointing this out. There are several differences between the two studies, ours and theirs:

      a. While the earlier study by Schallmo et al. (2018) employed 3T MRS, we utilize 7T MRS, enhancing our ability to detect and measure GABA with greater accuracy.

      b. Schallmo elife 2018 choose to use the bilateral hMT+ as the MRS measurement region while we use the left hMT+. The reason why we focus on left hMT+ are described in review 1. (6). Briefly, use of left MT/V5 as a target was motivated by studies demonstrating that left MT/V5 TMS is more effective at causing perceptual effects (Tadin et al., 2011).

      c. The resolution of MRS sequence in Schallmo elife 2018 is 3 cm isotropic voxel, while we apply 2 cm isotropic voxel. This helps us more precisely locate hMT+ and exclude more white matter signal.

      (3) Table 1 and Table Supplementary 1-3 contain many correlation results. But what are the main points of these values? Which values do the authors want to highlight? Why are only p-values shown with significance symbols in Table Supplementary 2?

      (3.1) what are the main points of these values?

      Thank you to the reviewer for pointing this out. These correlations represent the relationship between behavior task (SI/BDT) and resting-state functional connectivity. It indicates that left hMT+ is involved in the efficient information integration network when it comes to the BDT task. In addition, left hMT+’s surround suppression is involved in several hMT+ - frontal connectivity. Furthermore, the overlapping regions between two tasks indicate a shared underlying mechanism.

      (3.2) Which values do the authors want to highlight?

      Table 1 and Table Supplementary 1-3 present the preliminary analysis results for Table 2 and Table Supplementary 4-6. So, we generally report all value. Conversely, in the Table 2 and Table Supplementary 4-6, we highlight (bold font) indicating the significant correlations survived from multi correlation correction.

      (3.3) Why are only p-values shown with significance symbols in Table Supplementary 2?

      Thank you for pointing this out, it is a mistake. We have revised it and delete the significance symbols.

      (4) Line 27, it is unclear to me what is "the canonical theory".

      We thank the reviewer for pointing this out. We have revised “the canonical theory" to “the prevailing opinion”.

      (5) Throughout the paper, the authors use "MT+", I would suggest using "hMT+" to indicate the human MT complex, and to be consistent with the human fMRI literature.

      We thank the reviewer for pointing this out. We have revised them and used "hMT+" to be consistent with the human fMRI literature.

      (6) At the beginning of the results section, I suggest including the total number of subjects. It is confusing what "31/36 in MT+, and 28/36 in V1" means.

      We thank the reviewer for pointing this out. We have included the total number of subjects in the beginning of result section.

      (7) Line 138, "This finding supports the hypothesis that motion perception is associated with neural activity in MT+ area". This sentence is strange because it is a well-established finding in numerous human fMRI papers. I think the authors should be more specific about what this finding implies.

      We thank the reviewer for pointing this out. We have deleted the inappropriate sentence "This finding supports the hypothesis that motion perception is associated with neural activity in MT+ area".

      (8) There are no unit labels for all x- and y-axies in Figure 1. I only see the unit for Conc is mmol per kg wet weight.

      We thank the reviewer for pointing this out. Figure 1 is a schematic and workflow chart, so labels for x- and y-axes are not needed. I believe this confusion might pertain to Figure 3. In Figures 3a and 3b, the MRS spectrum does not have a standard y-axis unit as it varies based on the individual physical conditions of the scanner; it is widely accepted that no y-axis unit is used. While the x-axis unit is ppm, which indicate the chemical shift of different metabolites. In Figure 3c, the BDT represents IQ scores, which do not have a standard unit. Similarly, in Figures 3d and 3e, the Suppression Index does not have a standard unit.

      (9) Although the correlations are not significant in Figure Supplement 2&3, please also include the correlation line, 95% confidence interval, and report the r values and p values (i.e., similar format as in Figure 1C).

      We thank the reviewer for pointing this out. We have revised them.

      (10) There is no need to separate different correlation figures into Figure Supplementary 1-4. They can be combined into the same figure.

      We thank the reviewer for the suggestion. However, each correlation figure in the supplementary figures has its own specific topic and conclusion. The correlation figures in Supplementary Figure 1 indicate that GABA in V1 does not show any correlation with BDT and SI, illustrating that inhibition in V1 is unrelated to both 3D visuo-spatial intelligence and motion suppression processing. The correlations in Supplementary Figure 2 indicate that the excitation mechanism, represented by Glutamate concentration, does not contribute to 3D visuo-spatial intelligence in either hMT+ or V1. Supplementary Figure 3 validates our MRS measurements. Supplementary Figure 4 addresses potential concerns regarding the impact of outliers on correlation significance. Even after excluding two “outliers” from Figures 3d and 3e, the correlation results remain stable.

      (11) Line 213, as far as I know, the study (Melnick et al., 2013) is a psychophysical study and did not provide evidence that the spatial suppression effect is associated with MT+.

      We thank the reviewer for pointing this out. It was a mistake to use this reference, and we have revised it accordingly.

      (12) At the beginning of the results, I suggest providing more details about the motion discrimination tasks and the measurement of the BDT.

      We thank the reviewer for pointing this out. We have included some brief description of task at the beginning of the result section.

      (13) Please include the absolute duration thresholds of the small and large sizes of all subjects in Figure 1.

      We thank the reviewer for the suggestion. We have included these results in Figure 3.

      (14) Figure 5 is too small. The items in plot a and b can be barely visible.

      We thank the reviewer for pointing this out. We increase the size and resolution of Figure 5.

      Reviewer #2 (Recommendations For The Authors):

      Recommendations for improving the writing and presentation.

      I highly recommend editing the manuscript for readability and the use of the English language. I had significant difficulties following the rationale of the research due to issues with the way language was used.

      We thank the reviewer for pointing this out. We apologize for any shortcomings in our initial presentation. We have invited a native English speaker to revise our manuscript.

    1. Eventually, there will be different ways of paying for different levels of quality. But today there some things we can do to make better use of the bandwidth we have, such as using compression and enabling many overlapping asynchronous requests. There is also the ability to guess ahead and push out what a user may want next, so that the user does not have to request and then wait. Taken to one extreme, this becomes subscription-based distribution, which works more like email or newsgroups. One crazy thing is that the user has to decide whether to use mailing lists, newsgroups, or the Web to publish something. The best choice depends on the demand and the readership pattern. A mistake can be costly. Today, it is not always easy for a person to anticipate the demand for a page. For example, the pictures of the Schoemaker-Levy comet hitting Jupiter taken on a mountain top and just put on the nearest Mac server or the decision Judge Zobel put onto the Web - both these generated so much demand that their servers were swamped, and in fact, these items would have been better delivered as messages via newsgroups. It would be better if the ‘system’, the collaborating servers and clients together, could adapt to differing demands, and use pre-emptive or reactive retrieval as necessary.

      It's hard to make sense of these comments in light of TBL's frequent claims that the Web is foremost about URLs. (Indeed, he starts out this piece describing the Web as a universal information space.) It can really only be reconciled if you ignore that and understand "the Web" here to mean HTML over HTTP.

      (In any case, the remarks and specific examples are now pretty stale and out of date.)

    1. Note: This preprint has been reviewed by subject experts for Review Commons. Content has not been altered except for formatting.

      Learn more at Review Commons


      Referee #2

      Evidence, reproducibility and clarity

      Koszela et al. have submitted this manuscript demonstrating the molecular mechanism of interaction between Parkin and one of its known substrates, Miro1. While the interaction and ubiquitination of Miro1 by Parkin (and it's role in mitochondrial quality control) has been known since 2011, as demonstrated by the Schwarz group and others, the mechanism of action has remained unknown. The ability of Parkin to ubiquitinate multiple proteins upon mitochondrial damage has indeed led many groups to speculate that Parkin is a promiscuous E3 ligase upon activation; this manuscript tries to provide a rationale for the interaction with one of its known substrates through a combination of biochemical and biophysical studies.

      The authors demonstrate that Miro1 is efficiently ubiquitinated in in vitro biochemical assays in comparison to a few mitochondrial and non-mitochondrial proteins in an attempt to show that Miro1 is a preferred substrate for Parkin. Cross-linking coupled with mass spectrometry, SAXS and NMR experiments were used to provide compelling evidence for a direct and specific interaction between Parkin and Miro1. Molecular modelling using Colabfold and biochemical assays with mutants of the proposed interaction site were then used to provide further proof for the specificity of the interaction. This interaction is shown to occur between the conserved a.a.115-122 (referred to in this study as STR; located in the linker connecting the Ubl to RING0) and the EF domain of Miro1. Interestingly, the authors show that peptides corresponding to 115-122 competitively inhibit ubiquitination of Miro1 by Parkin. Overall, this article constitutes an important addition to our understanding of Parkin's mechanism of action. However, some of the key claims remain unsubstantiated, as described below.

      Major issues:

      1. In line 151 the authors claim, 'these data strongly support the hypothesis that Miro1 is the preferred substrate of pParkin...'. Arguably, the biggest issue with this study is the lack of substantial proof that Miro1 is the preferred parkin substrate in a cellular or physiological context. This claim cannot be made based on a biochemical assay with three other proteins. The Harper group has performed in-depth proteomics studies on the kinetics of Parkin-mediated ubiquitination and proposed that VDACs and Mfn2 (among a few others) are most efficiently ubiquitinated upon mitochondrial damage in induced neurons (Ordureau et al, 2018,2020). Interestingly, neither of these papers have been mentioned by the authors in this manuscript. The Trempe group has shown that Mfn2 is efficiently targeted by Parkin through mitochondrial reconstitution assays and proximity ligation assays (Vranas et al, 2022). The authors need to substantiate their claim through cellular or mitochondrial assays to prove that Miro1 is the preferred physiological substrate of Parkin. Cellular experiments also account for cellular abundance and proximity of Parkin to the substrate, which is not possible in biochemical assays of the kind presented here. In the absence of strong experimental proof for this claim, these claims should be tampered down to Miro1 being "the preferred substrate compared to the other proteins in this assay", and the manuscript should focus more on the molecular mechanism of interaction between Miro1 and Parkin.
      2. In addition to the point above, the authors do not describe the rationale for specifically choosing Mfn1 and MitoNEET for their comparison with Miro1 as substrates. Interestingly, Miro1, MitoNEET and Mfn1 are not among the most efficiently ubiquitinated substrates of Parkin (Ordureau et al, 2018). Additionally, the authors have used a construct of Mfn1 that lacks the full HR1 domain for their assays. Previously, it has been shown that the HR1 of mitofusins is targeted by Parkin (McLelland et al. 2018). Can the authors prove that their Mfn1 construct is as efficiently ubiquitinated as full-length Mfn1 by Parkin? If it is not possible to obtain soluble full-length Mfn1 or other membrane proteins for these assays, then I strongly recommend the authors should perform mitochondrial reconstitution assays as others have performed previously (Vranas et al, 2022) and use this opportunity to also report the ubiquitination kinetics of multiple mitochondrial substrates compared to Miro1 to make a more compelling case for substrate preference.
      3. The authors show that both pParkin-Miro1 and Parkin-Miro1 complexes can be captured by chemical cross-linking. It is well-established in the field that pUbl binds to RING0 (Gladkova et al, 2018) (Sauve et al, 2018) while non-phosphorylated Ubl binds RING1 (Trempe et al, 2013). The Komander group has also shown that the ACT (adjacent to the STR) element binds RING2 in the activated Parkin structure (Gladkova et al, 2018). This suggests that STR could occupy different positions in the Parkin and pParkin. The authors have only reported the cross-link/MS data and model of the Parkin-Miro1 complex. Arguably, the pParkin-Miro1 data is just as, if not more, relevant given that pParkin represents the activated form the ligase. The authors need to robustly establish that Miro1 binds to the STR element in both cases by demonstrating the following:

      A. Mass spectrometry data from cross-linked pParkin-Miro1 complex suggesting the same interaction site.

      B. Colabfold modelling with the pParkin structure to show that Miro1 would bind to the same element. 4. Does Parkin only bind to Miro1, or can it bind to Miro2 as well? Are there differences between the binding site and Ub target sites between the two proteins? The author should also show experimentally if both proteins get ubiquitinated as efficiently by Parkin and if the STR element is involved in recognizing both proteins. Interestingly, the Harper group reports that Miro2 gets more efficiently ubiquitinated than Miro1 (Ordureau et al, 2018). 5. In Figure 5D, the level of unmodified Miro1 seems to be similar in assays with WT or I122Y Parkin, though the former seems to form longer chains while the latter forms shorter chains. Is there an explanation for this? Perhaps, the authors need to perform this assay at shorter time points to show that there is more unmodified Miro1 remaining when treated with I122Y Parkin (and similarly for the L221R mutant of Miro1)? Also, why is the effect of Miro1 L221R and Parkin I122Y not additive?

      Minor comments:

      1. The authors should report the full cross-linking/MS data report from Merox including the full peptide table and decoy analysis report.
      2. The authors should report statistics for the fit of the Colabfold model to the experimental SAXS curve.
      3. Why is the Parkin-Miro1 interaction only captured by NMR and not by ITC? The authors should at least attempt to show the interaction of the STR peptide with Miro1 by an orthogonal technique like ITC.
      4. The authors should report the NMR line broadening data quantitatively i.e. reporting the reduction in signal intensity for the peaks upon peptide Miro1 binding to quantitatively demonstrate that the 115-122 peak intensity reduction is more significant than other regions.
      5. Figure 4 (structure figure) and B (PAE plot) should be annotated with the names of domains and elements in Parkin and Miro1 to make these figures clearer and more informative.

      Referees cross-commenting

      I am in agreement with reviewers 1 and 2. Both of them raise valid and interesting points in their reviews.

      Specifically, I would like to highlight the following:

      1. Reviewer 1 makes a very good point (5/6) highlighting that L119A does not impair Parkin recruitment in the previously reported study. I second this concern and believe that the authors need to re-frame their discussion and make it much more nuanced with regards to the role of Miro1-Parkin interaction in mitophagy (if any at all). Additionally, the authors should also note that previous studies in the field from the Youle group (Narendra et al, 2008) and multiple other groups have shown a complete absence of Parkin recruitment to healthy mitochondria. Parkin recruitment to healthy mitochondria hence remains a controversial idea at best, with no evidence for it outside of Parkin overexpression systems (Safiulina et al, 2018) which can also lead to artifacts. The discussion should take all major studies/observations into account to propose a more nuanced picture of the role of Parkin-Miro1 interaction. Perhaps, this interaction plays more of a role in mitochondrial quarantine (Wang et al. 2011) as suggested by the Schwarz group than in Parkin recruitment?
      2. Reviewer 3 raises a valid concern about the lack of quantification in ubiquitination assays and alludes to the difficulty in visualizing ubiquitination of multiple proteins. That was a concern I also had but did not include in my review. Perhaps, the authors should also show western blots for each of the protein (in a time course experiment) demonstrating the difference in ubiquitination kinetics of each of proteins instead of busy SDS-PAGE gels for the assay.

      Significance

      The key strength of this study is the strong biophysical evidence of a direct interaction between Parkin and Miro1 and the discovery of the Miro1 binding site on Parkin. The biophysical and biochemical experiments in this study have been well-designed and executed. The evidence for a specific interaction between Parkin and Miro1 has been provided through multiple approaches. The authors should be commended for this effort. The biggest limitation of this study is the lack of proof that Miro1 is the preferred Parkin substrate in a cellular/physiological context since in biochemical assays Parkin can ubiquitinate multiple proteins non-specifically. Substrate preference claims need to be established in more physiologically relevant experimental settings.

      Overall, the study represents a mechanistic advance in terms of our understanding of the interaction between Parkin and one of its substrates i.e. Miro1, showing that Parkin can indeed specifically bind its substrates before targeting them for ubiquitination. This might also inspire others to investigate the molecular mechanism of action of Parkin with other substrates. This paper would likely appeal specialized audiences i.e. biochemists and structural biologists studying Parkin in mitochondrial quality control.

      Reviewer expertise: Expert biochemist and biophysicist with a number highly cited works in the field of mitochondrial quality control and Parkin.

    1. Welcome back.

      I spent the last few lessons going through DNS, helping you, I hope, understand how the system works at an architectural level. In this lesson, I want to finish off and talk about the types of records which can be stored in DNS, and I'll try to keep it quick, so let's get started.

      The first record type that I want to touch on are nameserver records or NS records. I've mentioned these in the previous lessons in this section on DNS. These are the record types which allow delegation to occur in DNS. So we've got the dot com zone, and that's managed by Verisign. This zone will have multiple nameserver records inside it for amazon.com. These nameserver records are how the dot com delegation happens for amazon.com, and they point at servers managed by the amazon.com team. These servers host the amazon.com zone. Inside this one are DNS records such as www, which is how you can access those records as part of DNS.

      Now, of course, the same is true on the other side. The root zone has delegated management of dot com by having nameservers in the root zone point at the servers that host the dot com zone. So nameserver records are how delegation works end-to-end in DNS. Nameservers are hugely important.

      Next up, we have a pair of record types that you will use a lot more often in DNS, and they're A records or AAAA records, and they actually do the same thing. Given a DNS zone, in this example, google.com, these types of records map host names to IP addresses. The difference is the type of IP address. For a given host, let's say www, an A record maps this onto an IP version four address. An AAAA record type is the same, but this maps the host onto an IP version six address. Generally, as an admin or a solutions architect, you will normally create two records with the same name. One will be an A record, and one will be an AAAA record. The client operating system and DNS software on that client can then pick the correct type of address that it wants, either AAAA, if it's capable of IP version six, or just a normal A record, if it's not capable of version six.

      Now next up is the CNAME record type, which stands for canonical name. For a given zone, the CNAME record type lets you create the equivalent of DNS shortcuts, so host to host records. Let's say that we have an A record called server, which points at an IP version four address. It's fairly common that a given server performs multiple tasks. Maybe in this case, it provides ftp, mail, and web services. Creating three CNAMEs and pointing them all at the A server record means that they will all resolve to the same IP version four address. CNAMEs are used to reduce admin overhead. In this case, if the IP version four address of the server changes, it's just the single record to update, the A record, because the three CNAMEs reference that A record, they'll automatically get updated. Now, CNAMEs cannot point directly at an IP address, only other names, and you can expect to see that feature in the exam as a trick question.

      Next is the MX record type, and this is hugely important for how the internet works, specifically how email on the internet works. Imagine if you're using your laptop via your email server and you want to send an email to hi@google.com. MX records are used as part of this process. Your email server needs to know which server to pass the email onto. So we start with the google.com zone. Inside this zone, we have an A record with the name mail, and this is pointing at an IP address. Now it's important to know from the offset that this could be called rabbits or apple or fluffy; the name isn't important to how email works using MX records. In this case, the A record is just called mail, but it doesn't matter.

      Now also inside the google.com zone is a collection of MX records, in this example, two records. MX records have two main parts, a priority and a value, and I'll revisit the priority soon. For now, let's focus on the values. The value can be just a host, as with the top example. So mail here is just mail. That's just a host. If it's just a host and we can tell that by the fact that it's got no dot on the right, it's assumed to be part of the same zone that it's in. So mail here actually means mail.google.com. It's the mail host inside the google.com zone. If you include a dot on the right, this means it's a fully qualified domain name. And so it can either point to the host inside the same zone or something outside that zone, maybe Office 365 if Google decided Microsoft's mail product was better.

      The way that MX records are used is that our email server looks at the two addresses on the mail, so hi@google.com, and it focuses on the domain, so google.com. It then does an MX query using DNS on google.com. This is the same process as any other record type, so it talks to the root first, then dot com, then google.com, and then it retrieves any MX records. In this case, two different records. Now, this is where the priority value is used to choose which record to use. Lower values for the priority field are actually higher priority. So in this example, mail is used first and then mail.other.domain is only used if mail isn't functional. If the priority is the same, then any of them could be selected. Whichever is used, the server gets the result of the query back and it uses this to connect to the mail server for google.com via SMTP and it uses this protocol to deliver the mail. So in summary, an MX record is how a server can find the mail server for a specific domain. MX records are used constantly. Whenever you send an email to a domain, the server that is sending the email on your behalf is using DNS to do an MX lookup and locate the mail server to use.

      The last record type that I want to talk about is a TXT record, also known as a text record. Text records allow you to add arbitrary text to a domain. It's a way in which the DNS system can provide additional functionality. One common usage for a TXT record type is to prove domain ownership. Let's say for the Animals for Life domain, we want to add it to an email system, maybe Google Mail or Office 365 or Amazon WorkMail. Whatever system we use to host our email might ask us to add a text record to the domain, containing a certain piece of text data. So let's say that the random text that we need to add is "cats are the best." Then our administrator would add a record inside this domain with that text data. And once our admin has done that, the external party, so the Google email system, would query that text data, make sure that it matches the value that they're expecting. And if it does, that would prove that we own that domain and we can manage it. So text records are fairly important in proving domain ownership, and that's one of the most common use cases that you will use the text record type for. There are other uses for the text record type. It can be used to fight spam. So you can add certain information to a domain indicating which entities are authorized to send email on your behalf. If any email servers receive email from any other servers, then that's a good indication that that email is spam and not authorized.

      So those are the record types that I want to cover. But there's one more concept that I need to discuss before we finish up. And that is DNS TTL or Time To Live. A TTL value is something that can be set on DNS records. It's a numeric value in seconds. Let's look at a visual example. We have a client looking to connect to amazon.com. And so it queries DNS using a resolver server that's hosted at its internet provider. That resolver server talks to the DNS root, which points at the dot com registry authoritative servers. And so the resolver queries those servers. Those authoritative servers for dot com provide the nameservers at the amazon.com zone, and so it goes ahead and queries that. That server hosts and is authoritative for the amazon.com zone, which has a record for www. And so it uses this record to get the IP address and connect to the server.

      This process takes time. This walking the tree process, talking to the root, and then all of the levels to get the eventual result that you need, it is a lengthy process. Getting a result from the authoritative source, so the source that is trusted by DNS, this is known as an authoritative answer. So you get an authoritative answer by talking to a nameserver, which is authoritative for that particular domain. So if I query the nameserver for amazon.com and I'm querying the www record in amazon.com, then I get back what's known as an authoritative answer. And that is always preferred because it's always going to be accurate. It's the single source of truth.

      But using TTL values, the administrator of amazon.com can indicate to others how long records can be cached for, what amount of time is appropriate. In this example, because the admin of amazon.com has set a 3,600 TTL value, which is in seconds, it means that the results of the query are stored at the resolver server for 3,600 seconds, which is one hour. If another client queries the same thing, which is pretty likely for amazon.com, then they will get back a non-authoritative answer. But that answer will be retrieved immediately because it's cached on the resolver server. The resolver server, remember, is hosted probably at our internet provider, and so it's much quicker to access that data.

      So non-authoritative answers are often the same as authoritative answers. Normally things in DNS don't change, and when they don't change, non-authoritative and authoritative is the same thing. But TTL is important for when things change. If you migrate your email service and you have a high TTL value on your MX record, and you change to a provider with a different IP address, then email delivery might be delayed because old IP addresses for those MX records will be cached and they will be used. TTLs are a balance. Low values mean more queries against your nameservers. High values mean fewer queries, but also less control if you need to change records. You can change TTL values before projects and upgrades or you can leave them permanently low. Also, keep in mind that the resolver should obey TTL values, but that's not always the case. It could ignore them. That configuration can be changed locally by the admin at the resolver server. DNS is often the cause of project failures because of TTL values. If you're doing any work that involves changing any DNS records, it's always recommended to lower the TTL value well in advance of the work, sometimes days or weeks in advance, and this will make sure that you have fewer caching issues when you finally do change those records.

      Okay, that's it. That's everything I wanted to cover in this lesson. I've covered the different DNS record types, as well as introduced you to the TTL concept, which is essential to understand if you want to avoid any DNS-related problems. Thanks for listening. Go ahead, complete this video and when you're ready, join me in the next.

    1. Welcome back and in this demo lesson I'm going to step through how you can register a domain using Route 53. Now this is an optional step within the course. Worst case you should know how to perform the domain registration process within AWS and optionally you can use this domain within certain demos within the course to get a more real-world like experience.

      To get started, as always, just make sure that you're logged in to the IAM admin user of the general AWS account which is the management account of the organization. Now make sure that you have the Northern Virginia region selected. While Route 53 is a global service, I want you to get into the habit of using the Northern Virginia region. Now we're going to be using the Route 53 product, so click in the search box at the top of the screen, type Route 53 and then click to move to the Route 53 console.

      Now Route 53, at least in the context of this demo lesson, has two major areas. First is hosted zones and this is where you create or manage DNS zones within the product. Now DNS zones, as you'll learn elsewhere in the course, you can think of as databases which store your DNS records. When you create a hosted zone within Route 53, Route 53 will allocate four name servers to host this hosted zone. And that's important, you need to understand that every time you create a new hosted zone, Route 53 will allocate four different name servers to host that zone. Now the second area of Route 53 is registered domains, and it's in the registered domains area of the console where you can register a domain or transfer a domain in to Route 53.

      Now we're going to register a domain, but before we do that, if you do see any notifications about trying out new versions of the console, then go ahead and click to try out that new version. Where possible, I always like to teach using the latest version of the console UI because it's going to be what you'll be using long-term. So in my case, I'm going to go ahead and click on, try out the new console, depending on when you're doing this demo, you may see this or not. In either case, you want to be using this version of the console UI. So if you are going to register a domain for this course, then you need to go ahead and click register domains.

      The first step is to type the domain that you want into this box. Now, a case study that I use throughout the course is animals for life. So I'm going to go ahead and register a domain related to this case study. So if I type animalsforlive.com and press enter, it will search for the domain and tell us whether it's available. In this case, animalsforlive.com is not available. It's already been registered. In my case, I'm going to use an alternative, so I'm going to try and register animalsforlive.io. Now, I/O domains are one of the most expensive, so if you are registering a domain yourself, I would tend to advise you to look for one of the cheaper ones. I'm going to register this one and it is available.

      Once I've verified that it is available and it's the one I want, we're gonna go ahead and click on select. We can verify the price of this domain for one year, in this case it's 71 US dollars, and then go ahead and click on proceed to check out. Now it's here where you can specify a duration for the domain registration. You can use the default of one year, or alternatively you can go ahead and pick a longer registration period. For this domain I'm going to choose one year and then you can choose whether you want to auto renew the domain after that initial period. In my case I'm going to leave this selected. You'll see a subtotal at the price and then you can click next to move on to the next step.

      Now at this point you need to specify the contact type. In most cases you'll be putting a person or a company but there's also association, public body or reseller. You need to go ahead and fill in all of these details and they do need to be valid details, that's really important. If you are worried about privacy, most domains will allow you to turn on privacy protection, so any details that you enter here cannot be seen externally. Now obviously to keep my privacy intact, I'm going to go ahead and fill in all of these details and I'm going to hide the specifics and once I've entered them all, I'm going to go ahead and click on 'Next' and you should do the same. Again I've hidden my details on the bottom of the screen.

      Route 53 does tell you that in addition to the domain registration cost there is a monthly cost for the hosted zone which will be created as part of this registration. So there is a small monthly cost for every hosted zone which you have hosted using Route 53 and every domain that you have will need one hosted zone. So I'm going to scroll down. Everything looks good, you'll need to agree to the terms and conditions and then click on submit. Now at this point the domain is registering and it will take some time to complete. You may receive a registration email which may include something that you need to do, clicking on a link or some other form of identity verification. You might not get that, but if you do get it, it's important that you do follow all of the steps contained within that email. And if you don't receive an email, you should check your spam folder, because if there are any actions to perform and you don't, it could result in the domain being disabled.

      You can see the status of the domain registration by clicking on "requests" directly below "registered domains". The status will initially be listed as "in progress", and we need this to change to "successful". So pause the video, wait for this status to change, and then you're good to continue. Welcome back, in my case this took about 20 minutes to complete, but as you can see my domain is now registered. So if we go to registered domains you'll be able to see the domain name listed together with the expiration date, the auto renew status, and the status of the transfer lock. Now transfer lock is a security feature, it means the domain cannot be transferred away from route 53 without you disabling this lock.

      Now we're able to see additional details on the domain if we click on the domain name. Now obviously I've hidden my contact information. If you click on the DNSsecKeys tab then it's here where you can configure DNSsec on the domain. We won't be doing anything with that at this stage. One of the important points I want to draw your attention to is the name servers. So I've registered animalsforlife.io and it's these name servers that will be entered into the Animals for Life record within the .io top level domain zone. So these servers are the ones that the DNS system will point at. These currently are set to four Route 53 name servers. And because we've registered the domain inside Route 53, this process is automatic. So a hosted zone is created, four name servers are allocated to host this hosted zone And then those four name servers are entered into our domain records in our top level domain zone.

      This process end-to-end is all automatic. So the four name servers for the animalsforlife.io hosted zone. These are entered into the animalsforlife.io record within the .io top level domain zone. It's all automatic. So if we move to the hosted zone area of the console and then go inside AnimalsForLife.io and then expand the hosted zone details at the top These are the four name servers which are hosting this hosted zone And if you're paying attention You'll note these are the same four servers that are contained within the registered domains Area of the console and these are the same four servers which have been entered into the .io top level domain zone. Now if you ever delete and then recreate a hosted zone It's going to be allocated with four brand new name servers. These name servers will be different than the name servers for the zone which you deleted So if you delete and recreate a hosted zone You'll be given four brand new name servers. In order to stop any DNS problems you'll need to take these brand new name servers and update the items within the registered domains area of the console but again because you've registered the domain within route 53 this process has been handled for you end to end you won't need to worry about any of this unless you delete and recreate the host of zone.

      Now that's everything you need to do at this point if you followed this process throughout this demo lesson you now have an operational domain within the global DNS infrastructure that's manageable within Route 53. Now as I mentioned earlier this is an optional step for the course if you do have a domain registered then you will have the opportunity to use it within various demo lessons within the course. If you don't, don't worry, none of this is mandatory you can do the rest of the course without having a domain. At this point though that is everything I wanted you to do in this demo lesson. Go ahead and complete the video and when you're ready I'll look forward to you joining me in the next.

    1. Welcome back. And now that I've talked about the fundamentals of DNS from an abstract perspective, I want to bring this back to an AWS focus and talk about Route 53, which is AWS's managed DNS product.

      Okay, let's jump in and get started with a high level product basics, and then I'll talk about the architecture. Route 53 provides two main services. First, it's a service in AWS, which allows you to register domains. And second, it can host zone files for you on managed name servers, which it provides. Now Route 53 is a global service with a single database. It's one of very few AWS services which operates as a single global service. And as such, you don't need to pick a region when using it from the console UI. The data that Route 53 stores or manages is distributed globally as a single set and it's replicated between regions. And so it's a globally resilient service. Route 53 can tolerate the failure of one or more regions and continue to operate without any problems. Now it's one of the most important AWS products. It needs to be able to scale, stay highly performant, whilst remaining reliable, and continue working through failure.

      So let's look at exactly how Route 53 is architected and exactly what it does to provide these two main services. So the first service that I mentioned at the start of this lesson is that Route 53 allows you to register domains. And to do that, it has relationships with all of the major domain registries. Remember from the last lesson that these are the companies which manage the top level domains. They've been delegated this ability by IANA who manage the root zone for DNS. Now these registries, each manage one specific zone. One of them manages the .com zone and/or the .net zone, and another the .io zone, and so on.

      In the next lesson, I'll be demoing how to register a domain that I'll be using for the course scenario. And that domain will be a .org domain. And so one of these relationships is with the .org registry, an organization called PIR. Now, when a domain is registered, a few things happen. First, Route 53 checks with the registry for that top level domain if the domain is available. For this example to keep it simple, let's just assume it is. Then Route 53 creates a zone file for the domain being registered. And remember a zone file is just a database which contains all of the DNS information for a particular domain. In this case, animals4life.org. As well as creating the zone file, Route 53 also allocates name service for this zone. So these are servers which Route 53 creates and manages which are distributed globally and there are generally four of these for one individual zone.

      So it takes this zone file that it's created. And this is known as a hosted zone, using Route 53 terminology, and it puts that zone file onto these four managed name servers. And then as part of registering the domain it communicates with the .org registry. And this is PIR in this case, and liaising with that registry, it adds these name server records into the zone file for the .org top level domain. And the way that it does this is it uses name server records. So these name server records are how PIR delegate the admin of the domain tools. By adding the name server records to the org zone, they indicate that our four name servers are all authoritative for the domain. And that's how a domain is registered using Route 53.

      It's not a complicated process when you simplify it right down. It's simply the process of creating a zone file, creating a number of managed name servers, putting that zone file on those servers, and then liaising with the registry for the top level domain, and getting a name server records added to the top level domain zone, which point back at these servers. Remember, DNS is just a system of delegation.

      So next, let's quickly take a look at zones inside Route 53. So Route 53 provides DNS zones as well as hosting for those zones. It's basically DNS as a service. So it lets you create a manage zone files. And these zone files are called hosted zones in Route 53 terminology, because they're hosted on AWS managed name servers. So when a hosted zone is created, a number of servers are allocated and linked to that hosted zone. So they're essentially one and the same. From Route 53's perspective, every hosted zone also has a number of allocated managed name servers. Now a hosted zone can be public, which means that the data is accessible on the public internet. The name servers for a public hosted zone live logically in the AWS public zone. And this is accessible anywhere with the public internet connection. So they're part of the public DNS system.

      A hosted zone could also be private which means that it's linked to one or more VPCs and only accessible from within those VPCs. And you might use this type of zone if you want to host sensitive DNS records that you don't want to be publicly accessible. A hosted zone hosts DNS records, which I'll be talking about in an upcoming lesson in much more detail because there are many different types of records. Inside Route 53, you'll see records referred to as record sets. Now there is a tiny difference, but for now you can think of them as the same thing.

      Okay, so now it's time for a demo. I know that DNS has been a lot of theory. And so I wanted to show you a domain being registered and the domain that will be registered is the domain that I'll be using for the course scenario which is animals4life.org. So when you're ready to see that, go ahead, complete this video, and join me in the next.

    1. Welcome back and in this demo lesson you're going to get some experience interacting with CloudWatch. So you're going to create an EC2 instance, you're going to cause that instance to consume some CPU capacity and then you're going to monitor exactly how that looks within CloudWatch. Now to do this in your own environment you'll just need to make sure that you're logged into the general AWS account as the IAM admin user and as always make sure that you have the Northern Virginia region selected which is US-East-1. Once you've got those set correctly then click in the search box at the top and type EC2, find the EC2 service and then just go ahead and open that in a brand new tab.

      Now we're going to skip through the instance creation process because you've done that in a previous demo lesson. So just go ahead and click on instances and then Launch Instance. Under Name, I just want you to put CloudWatch Test as the instance name. Then scroll down and then under the Amazon Machine image to use, go ahead and select Amazon Linux. We're going to pick the Amazon Linux 2023 version, so that's the most recent version of this AMI. It should be listed as Free Tier Eligible, so just make sure that's the case. We'll leave the architecture set to 64-bit x86 and scroll down. It should already be set to an instance type which is free tier eligible, in my case t2.micro. We'll be connecting to this instance using ec2 instance connect so we won't be using an SSH key pair. So in this drop down just click and then say proceed without a key pair. We won't need one because we won't be connecting with a local SSH client. Scroll down further still and under Network Settings click on Edit and just make sure that the default VPC is selected. There should only be one in this list but just make sure that it's set as default. Under Subnet we can leave this as No Preference because we don't need to set one. We will need to make sure that Auto Assign Public IP is set to Enable.

      Under create security group for the name and for the description just go ahead and type CloudWatch SG so CloudWatch SG for both the security group name and the description now the default for security group rule should be fine because it allows SSH to connect from any source location and that's what we want scroll down further still and we'll be leaving storage as default remember this is set from the AMI that we pick. Now because this is a CloudWatch lesson, we're going to set something a little bit different. So expand Advanced Details and then scroll down and look for Detailed CloudWatch Monitoring. Now this does come at an additional cost, so you've got a couple of options. You can just watch me do this or you can do this demo without Detailed Monitoring enabled. And if you don't enable this, it will be entirely free, but you might need to wait a little bit longer for things to happen in the demo lesson so keep that in mind.

      What I'm going to do is I'm going to enable detailed CloudWatch monitoring and if we click on info here we can see some details about exactly what that does and we can also open this in a new tab and explore what additional charges apply if we want to enable it. Now in this case I'm going to enable it you don't have to it's not a huge charge but I think for me demoing this to you it's good that I enable it you don't have to you might just have to wait a little bit longer for things to happen in the demo. Now once all of that set just scroll all the way down to the bottom and go ahead and click launch instance. Now this might take a few minutes to create we're first waiting for this success dialog and once that shows we can go ahead and click on view all instances. Go ahead and click refresh until you see the instance it will start off in a pending state with nothing listed under status check. After a few moments this will change status we'll see that it's in a running state and then we need to wait for this to change to two of two status checks before we continue. So go ahead and pause the video wait for your status check to update and once it does we're good to continue.

      Okay so now this has changed to two out of two checks passed and that's good that's what we want so so it should display running on the instant state and then two out of two checks passed under status check. Once this is the case, go ahead and click in the search box at the top and just type CloudWatch, locate the CloudWatch service, and then open that in a brand new tab. This is the CloudWatch console, and it's here where we're going to create a CloudWatch alarm. Now if you see anything about a new UI or new features, you can just go ahead and close down that dialog. Once we're here, go ahead and click on Alarms on the left and then click on all alarms. This will show a list of all the alarms that you've configured within CloudWatch, and currently there aren't any. What we're going to do is to create an alarm. So click on create alarm, and then click on select metric. Once we're on this screen, scroll down, and we're going to be looking for an EC2 metric, because we need to find the CPU utilization metric, which is inside the EC2 namespace. In other words, it comes from the EC2 service. So go ahead and click on EC2, and then we're looking for per instance metrics. So click on per instance metrics, and this will show all of the EC2 instance metrics that we currently have. Now if I scroll through this list, what you'll see is that I have two different instance IDs, because I'm using this account to create all of these demo lessons. In my case, I see previous instances. Now if you're doing this in your account, if you go back to the EC2 Management Console, you can see your instance ID here. Just remember the last four digits of this instance ID, and then go back to the CloudWatch Console. If you have more than one instance listed in CloudWatch, look for the instance ID that ends with the four digits that you just noted down, and then from that list you need to identify CPU utilization. And so I'm going to check the box next to this metric. Now this is the metric that monitors, as the name suggests, CPU utilization on this specific instance ID, which is our CloudWatch test instance. If I scroll up, I'm able to see any data that's already been gathered for this specific instance. And as you can see, it's not a great deal at the moment because we've only just launched this instance. So I'm gonna go ahead and click on Select Metric, and then because we're creating an alarm, it's going to ask us for what metric and conditions we want to evaluate.

      So I'm going to scroll down, and under Conditions, I'm going to pick Static, because I want this alarm to go into an alarm state when something happens to the CPU utilization. So I'm going to ask CloudWatch that whenever the CPU utilization is greater or equal to a specific value than to go into an alarm state. So that value is going to be 15%. So whenever the CPU utilization on this EC2 instance is greater or equal to 15%, then this alarm will go into the alarm state. So I'm gonna go ahead and click on Next. Now you can set this up so that if this alarm goes into an alarm state, it can notify you using SNS. Now that's useful if this is in production usage, but in this case we're not using it in production, so I'm going to go ahead and click on remove. Scroll down to the bottom, there's also other things that you could pick, so you could do an auto scaling action, an EC2 action, or a systems manager action. But we're going to be talking about these in much more detail as we move through the course. For now we're going to keep this simple, it's just going to be a basic alarm which goes into an alarm state or not. So click on next and then under alarm name I'm going to put CloudWatch test and then high CPU and you should do the same. So type that, click on next, scroll down to the bottom and create that alarm.

      Now initially this alarm state will be insufficient data because CloudWatch hasn't yet gathered enough data on the CPU utilization to generate the state. That's fine because we've we've got another thing that we need to do first. So now move back to the EC2 console and we're going to connect into this instance using EC2 Instance Connect. Remember, that's the web-based way to get access to this instance. So over the top of the CloudWatch Test instance, right click and go to Connect. Make sure that EC2 Instance Connect is selected, so click that tab. You can leave everything as default and click on Connect and that will connect you to this EC2 instance. Now at this point, we need to install an application called stress on this EC2 instance. And stress is an application which will put artificial CPU load onto a system. And that's what we want to do in order to see how CloudWatch reacts. To install stress, we're going to run this command. And this next command will use the yum package manager to install the stress utility. So go ahead and run this command and then clear the screen again. Now the stress command can be run by typing stress and what we're going to do is do a double hyphen help just to get the help for this command. So what we're going to do is we're going to run stress and we're going to specify the number of CPUs to use and we want that number to be the same number of virtual CPUs that this instance has. Now a t2.micro has one virtual CPU and so the command that we need to run is stress space hyphen c space 1 and then space and then we're going to use hyphen t which is the timeout command and this specifies how long we want to run this for. So we're going to specify 3600 so hyphen t and then a space 3600 and this will run the stress for 3600 seconds and that's plenty for us to see how this affects the metrics which are being monitored by CloudWatch.

      Now what I want to do before we do that is go back to the CloudWatch console. You might need to refresh if you haven't seen the state update yet. In my case it's already showing as okay. So this means that it's now got access to some data. So click on this alarm and you'll be able to see that currently the CPU started off at very low levels and then it spiked up and potentially in my case that's because we've just installed some software. But note here this red line which indicates the alarm level for this alarm. So if the CPU utilisation, which is in blue, exceeds this red line then this alarm will move from OK to ALARM. And that's what we want to simulate. So go back to the instance and press Enter to run this stress command. And that's going to begin placing high levels of CPU load on this instance and what we'll see over the next few minutes is CloudWatch will detect this additional CPU load and it will cause this alarm to go from OK into an alarm state. So move back to the CloudWatch console and just keep hitting refresh until you see a change in the alarm state. Again this might take a few minutes. What I suggest you do is pause the video and wait for your alarm to change away from OK and then you're good to continue.

      Now in my case this only took a few minutes and as you can see the CPU load reported by this alarm in CloudWatch went from this value here and spiked all the way up to this value which is well above the 15% of the alarm threshold. So the alarm changed from OK to IN alarm based on this excessive CPU and if we keep monitoring this over time you'll see that this trend continues because this CPU is under extremely high load because it's been artificially simulated using the stress utility. Now if we go back to this EC2 instance and press ctrl and C at the same time this will exit out of the stress utility and at this point the artificial CPU load has been removed and the instance will gradually move back down to its normal levels which is very close to zero. So again what you'll see is this may take a few minutes to be reflected inside CloudWatch. So keep refreshing this once you've cancelled the stress utility and wait for the reported CPU utilization to move back down below the alarm value. Again that might take a few minutes so go ahead and pause the video and wait for this blue line to move back under the red line and once it does you should see that the alarm state changes from in alarm to OK again.

      In my case it took a few minutes for the blue line to move below the alarm threshold and then a few more minutes afterwards for the alarm to change from in alarm to OK. But as you can see at this point that's exactly what's happened once the CPU usage goes below the configured threshold value then the alarm changes back to an OK state. And at this point that's everything that I wanted to cover in this demo lesson on CloudWatch. CloudWatch is a topic that I'm going to be going into much more detail later on in the course. This has just been a really brief introduction to the product and how it interacts with EC2. Now at this point the only thing left is to clear up the account and put it back into the same state as it was at the start of this lesson. So to do that go ahead and click on All Alarms, select the CloudWatch Test High CPU Alarm that you created, click on the actions dropdown, select delete, and then confirm that deletion. Then go back to EC2, go to the instances overview, right click on the CloudWatch test instance, making sure that it is the correct instance, so CloudWatch test, and then select terminate instance and confirm that termination. Now that's going to move through a few states, it will start with shutting down, and you need to wait until that instance is in a terminated state. Go ahead and pause the video and wait for your instance to change into terminated.

      Okay so once your instance has terminated on the menu on the left scroll down go to security groups select the CloudWatch SG security group making sure that you do pick the correct one so CloudWatch SG click on actions scroll down delete security groups and click on delete and at that point the account is back in the same state as it was at the start of this demo lesson. So thanks for watching this video. I hope you gained some experience of the CloudWatch product and again we're going to be talking about it in much more detail later in the course. At this point though go ahead and complete this video and when you're ready I'll look forward to you joining me in the next.

    1. Welcome back. In this demo lesson, I want to quickly demonstrate how to use CloudFormation to create some simple resources. So before we start, just make sure you're logged in to the general AWS account and that you've got the Northern Virginia region selected. Once you've got that, just move across to the CloudFormation console.

      So this is the CloudFormation console, and as I discussed in the previous lesson, it works around the concepts of stacks and templates. To get started with CloudFormation, we need to create a stack. When you create a stack, you can use a sample template, and there are lots of different sample templates that AWS makes available. You can create a template in the Designer or upload a ready-made template, and that's what I'm going to do. Now, I've provided a template for you to use, linked to this lesson. So go ahead and click on that link to download the sample template file.

      Once you've downloaded it, you'll need to select 'Upload a template file' and then choose 'File'. Locate the template file that you just downloaded; it should be called 'ec2instance.yaml'. Select that and click on 'Open'. Whenever you upload a template to CloudFormation, it's actually uploading the template directly to an S3 bucket that it creates automatically. This is why, when you're using AWS, you may notice lots of buckets with the prefix CF that get created in a region automatically. You can always go ahead and delete these if you want to keep things tidy, but that's where they come from.

      Now, before we upload this, I want to move across to my code editor and step through exactly what this template does. The template uses three of the main components that I've talked about previously. The first one is parameters. There are two parameters for the template: latest AMI and SSH and web location. Let's quickly talk about the latest AMI ID because this is an important one. The type of this parameter is a special type that's actually a really useful feature. What this allows us to do is rather than having to explicitly provide an AMI ID, we can say that we want the latest AMI for a given distribution. In this case, I'm asking for the latest AMI ID for Amazon Linux 2023 in whichever region you apply this template in. By using this style of parameter, the latest AMI ID gets set to the AMI of the latest version of this operating system.

      The final parameter that this template uses is SSH and web location, which is where we can just specify an IP address range that we want to be able to access this EC2 instance. So that's parameters—nothing special, and you'll get more exposure to these as we go through the course. Now we've also got outputs, and outputs are things that are set when the template has been applied successfully. When a stack creates, when it finishes that process, it will have some outputs. I've created outputs so that we get the instance ID, the availability zone that the instance uses—remember EC2 is an AZ service. It’ll also provide the public DNS name for the instance, as well as the public IP address. The way that it sets those is by using what's known as a CloudFormation function.

      So this is ref, and this is going to reference another part of the CloudFormation template. In this case, it's going to reference a logical resource, the EC2 instance resource. Now, get attribute or get att is another function that's a more capable version of ref. With get attribute, you still refer to another thing inside the template, but you can pick from different data that that thing generates. An EC2 instance, by default, the default thing that you can reference is the instance ID, but it also provides additional information: which availability zone it's in, its DNS name, and its public IP. I’ll make sure to include a link in the lesson that details all of the resources that CloudFormation can create, as well as all of the outputs that they generate.

      The main component of course of this template is the resources component. It creates a number of resources. The bottom two, you don’t have to worry about for now. I’ve included them so I can demonstrate the Session Manager capability of AWS. I'll be talking about that much more later in the course, but what I'm doing is creating an instance role and an instance role profile. You won't know what these are yet, but I’ll be talking about them later in the course. For now, just ignore them. The main two components that we're creating are an EC2 instance and a security group for that instance.

      We’re creating a security group that allows two things into this instance: port 22, which is SSH, and port 80, which is HTTP. So it’s allowing two different types of traffic into whatever the security group is attached to. Then we’re creating the EC2 instance itself. We’ve got the EC2 instance, which is a logical resource, the type being AWS::EC2::Instance, and then the properties for that logical resource, such as the configuration for the instance. We’re setting the type and size of the instance, t2.micro, which will keep it inside the free tier. We’re setting the AMI image ID to use, and it's referencing the parameter, and if you recall, that automatically sets the latest AMI ID. We’re setting the security group, which is referencing the logical resource that we create below, so it creates this security group and then uses it on the instance. Finally, we’re setting the instance profile. Now, that’s related to these two things that I’m not talking about at the bottom. It just sets the instance profile, so it gives us the permission to use Session Manager, which I’ll demonstrate shortly after we implement this.

      There’s nothing too complex about that, and I promise you by the end of the course, and as you get more exposure to CloudFormation, this will make a lot more sense. For now, I just want to use it to illustrate the power of CloudFormation. So I’m going to move back to the console. Before I do this, I’m going to go to services and just open EC2 in a new tab. Once you’ve done that, return to CloudFormation and click on next. We’ll need to name the stack. I’m just going to call it CFN demo one for CloudFormation demo one. Here’s how the parameters are presented to us in the UI. The latest AMI ID is set by default to this value because, if we look at the parameters, it’s got this default value for this parameter. Then SSH and web location also has a default value which is set in the template, and that’s why it’s set in the UI. Leave these two values as default. Once you’ve done that, click on next.

      I’ll be talking more about all of these advanced options later on in the course when I talk about CloudFormation. For now, we’re not going to use any of these, so click on next. On this screen, we need to scroll down to the bottom and check this capabilities box. For certain resources that you can create within CloudFormation, CloudFormation views them as high-risk. In this case, we're creating an identity, an IAM role. Don't worry, I'll be talking a lot more about what an IAM role is in the next section of the course. Because it's an identity, because it's changing something that provides access to AWS, CloudFormation wants us to explicitly acknowledge that we’re to create this resource. So it’s prompting us for this capability to create this resource. Check this box, it’s fine, and then click on submit. The stack creation process will begin and the status will show create in progress.

      This process might take a few minutes. You’re able to click on refresh here, so this icon on the top right, and this will refresh the list of events. As CloudFormation is creating each physical resource that matches the logical resources in the template, it’s going to create a new event. For each resource, you’ll see a create in progress event when the creation process starts, and then you’ll see another one create complete when it creates successfully. If there are any errors in the template, you might see red text, which will tell you the nature of that error. But because this is a CloudFormation template that I’ve created, there’ll be no errors. After a number of minutes, the stack itself will move from Create in Progress to Create Complete.

      I refreshed a couple more times and we can see that the Session Manager instance profiles moved into the Create Complete status and straight after that it started to create the EC2 instance. We’ve got this additional event line saying Create in Progress, and the resource creation has been initiated. We’re almost at the end of the process now; the EC2 instance is going to be the last thing that the stack will create. At this point, just go ahead and pause the video and wait until both the EC2 instance and the stack itself move into Create Complete. Once both of those move into Create Complete, then you can resume the video and we’re good to continue.

      Another refresh, and we can see that the EC2 instance has now moved into a Create Complete status. Another refresh and the entire stack, CFN demo 1, is now in the create complete state, which means that the creation process has been completed and for every logical resource in the template, it’s created a physical resource. I can click on the outputs tab and see a list of all the outputs that are generated from the stack. You’ll note how they perfectly match the outputs that are listed inside the template. We’ve got instance ID, AZ, public DNS, and public IP. These are exactly the same as the outputs listed inside the CloudFormation template. You’ll see that these have corresponding values: the instance ID, the public DNS of the instance, and the public IP version 4 address of the instance.

      If I click on the resources tab, we’ll be able to see a list of the logical resources defined in the template, along with their corresponding physical resource IDs. For the EC2 instance logical resource, it’s created an instance with this ID. If you click on this physical ID, it will take you to the actual resource inside AWS, in this case, the EC2 instance. Now, before we look at this instance, I’m going to click back on CloudFormation and just click on the stacks clickable link at the top there. Note how I’ve got one stack, which is CFN demo one. I could actually go ahead and click on create stack and create stack with new resources and apply the same template again, and it would create another EC2 instance. That’s one of the powerful features of CloudFormation. You can use the same template and apply it multiple times to create the same set of consistent infrastructure.

      I could also take this template because it's portable, and because it automatically selects the AMI to use, I could apply it in a different region and it would have the same effect. But I’m not going to do that. I’m going to keep things simple for now and move back to the EC2 tab. Now, the one thing I want to demonstrate before I finish up with this lesson is Session Manager. This is an alternative to having to use the key pair and SSH to connect to the instance. What I’m able to do is right-click and hit Connect, and instead of using a standalone SSH client, I can select to use Session Manager. I’ll select that and hit Connect, and that will open a new tab and connect me to this instance without having to use that key pair.

      Now, it connects me using a different shell than I'm used to, so if I type bash, which is the shell that you normally have when you log into an EC2 instance, that should look familiar. I’m able to run normal Linux commands like df -k to list all of the different volumes on the server, or dmesg to get a list of informational outputs for the server. This particular one does need admin permission, so I’ll need to rerun this with sudo and then dmesg. These are all commands that I could run in just the same way if I was connected to the instance using an SSH client and the key pair. Session Manager is just a better way to do it, but it requires certain permissions to be given to the instance. That’s done with an instance role that I’ll be talking all about later on in the course. That is the reason why my CloudFormation template has these two logical resources, because these give the instance the permission to be able to be connected to using Session Manager. It makes it a lot easier to manage EC2 instances.

      So that’s been a demo of how easy it is to create an EC2 instance using CloudFormation. Throughout the course, we'll be using more and more complex examples of CloudFormation. I’ll be using that to show you how powerful the tool is. For now, it’s a really simple example, but it should show how much quicker it is to create this instance using CloudFormation than it was to do it manually. To finish up this lesson, I’m going to move back to the CloudFormation console. I’m going to select this CloudFormation stack and click on Delete. I need to confirm that I want to do this because it’s telling me that deleting this stack will delete all of the stack resources.

      What happens when I do this is that the stack deletes all of the logical resources that it has, and then it deletes all of the corresponding physical resources. This is another benefit of CloudFormation in that it cleans up after itself. If you create a stack and that creates resources, when you delete that stack, it cleans up by deleting those resources. So if I click on Delete Stack Now, which I will do, it starts a delete process, and that’s going to go ahead and remove the EC2 instance that it created. If I select this stack now, I can watch it do that. I can click on Events, and it will tell me exactly what it’s doing. It’s starting off by deleting the EC2 instance. If I move back to the EC2 console and just hit Refresh, we can see how the instance state has moved from running to shutting down.

      Eventually, once the shutdown is completed, it will terminate that instance. It’ll delete the storage, it will stop using the CPU and memory resources. At that point, the account won’t have any more charges. It wouldn’t have done anyway because this demo has been completely within the free tier allocation because I was using a t2.micro instance. But there we go. We can see the instance state has now moved to terminated. Go back to CloudFormation and just refresh this. We’ll see that it’s completed the deletion of all the other resources and then finished off by deleting the stack itself. So that’s the demonstration of CloudFormation. To reaffirm the benefits, it allows us to do automated, consistent provisioning. We can apply the same template and always get the same results. It’s completely automated, repeatable, and portable. Well-designed templates can be used in any AWS region. It’s just a tool that really does allow us to manage infrastructure effectively inside AWS.

    1. Welcome back and in this demo lesson, I just want you to get some experience working with S3.

      In this demo lesson you're going to create an S3 bucket which is going to be used for a campaign within the Animals for Life organization.

      You're going to get the chance to create the bucket, interact with the bucket, upload some objects to that bucket and then finally interact with those objects.

      Now to get started you'll need to make sure that you're logged in to the IAM

      admin user within the general AWS account. By this point in the course you

      should have a general account and a production account and you need to make

      sure that you're logged in to the general AWS account. As always make sure

      that you're also using the Northern Virginia region which is US-East-1.

      Now assuming that you do have that configuration next you need to move to

      the S3 console and there are a couple of ways that you can do that you can type

      S3 into this find services box if you've previously used the service it will be

      listed under the recently visited services and then finally at the top

      here you can click on the services drop-down and either type S3 into the

      all services box here or locate it in the list of services and click to move

      to the S3 console so I'm going to go ahead and type S3 and then click to move

      to the console. Now when you first arrive at the S3 console you'll be presented

      with a list of buckets within this AWS account. I want to draw specific

      attention to the fact that with S3 you do not have to choose a region with the

      region drop-down. When you create buckets within S3 you have to pick the region

      that that bucket is created in but because S3 uses a global namespace you

      don't have to select a region when using the console. So on this list you will see

      any buckets in any regions within this one single AWS account. You don't have to

      pick the region in advance. So let's go ahead and create an S3 bucket and to do

      that logically enough we click on create bucket. Now to create a bucket you need

      to specify a name and we're creating this bucket for a koala campaign for the

      Animals for Life organization. So we're going to start with Koala Campaign.

      Now because bucket names do need to be unique we can't just leave it at Koala Campaign

      we need to add some random numbers at the end. This is just to make sure that

      the name that you pick is different than the name that I pick and different than

      the name that every other student uses. So just put some numbers after this name.

      I'm going to pick 1-3-3-3-3-3-7. Now there are some rules around bucket

      naming names need to be between 3 and 63 characters they can only consist of

      lowercase letters numbers dots and hyphens they need to begin and end with a

      letter or number they can't be formatted like an IP address and they can't begin

      with X and N and of course they need to be entirely unique now there are some

      specific restrictions or specific rules for naming buckets if you want to use

      certain S3 features. Later in the course I'll be talking about static website

      hosting within S3 and I'll be showing you how you can use a custom domain name

      with an S3 bucket so you can get a domain name to host for example a blog

      or a static website and you can use S3 to host that website and if you want to

      do that then you need to name the bucket name the same as the DNS name that

      you'll be using to access that bucket. But at this point this is just an

      introductory demo so we can leave this as just a standard name. So use Koala

      campaign with some random numbers at the end and that should be good. Now when

      you're creating a bucket you need to specify a region and this is the region

      that this bucket will be placed in. Now I'm going to use US-East-1

      as a default throughout this course and so I do recommend that you pick that to

      create the bucket in. Now if you have any existing buckets within your account and

      you want to copy the settings from those buckets and this will of course just

      save you some time when setting up the bucket then you can click on choose

      bucket and copy the settings from another bucket in your account. Now

      because we're starting this from fresh and we don't have any existing buckets we

      can't use this option. So we need to scroll down and just review what options

      we have. For now we're going to skip past object ownership because this is a

      feature that I'll be discussing in much more detail later in the course. I can't

      really explain this until you have some experience of how the permissions model

      works with S3 so I'll be talking about it in the S3 section of the course. Now

      the first thing that you need to pick when you're creating buckets is this

      bucket settings for block public access. So all S3 buckets by default are

      private. Nobody has permissions to this bucket apart from the account that

      creates the bucket. So in this particular case we're creating it inside the

      general AWS account and so only the general AWS account and the account root

      user of that account have permissions. Now because we've granted the IAM admin

      user full admin permissions then it too has access to this bucket but by default

      nothing else can have access. Now you can make a bucket completely public. You can

      grant access for all users to that bucket including unauthenticated or

      anonymous users. Now that's a security risk because potentially you might have

      sensitive data within that bucket. This is a fail-safe. This means that even if

      you grant completely public access to a bucket then this will block that access.

      And I'm going to be talking about this in much more detail later in the course

      But you need to know that this exists if we untick this option for example

      Even though we are now not blocking all public access

      You still need to grant access to this bucket

      So all this option does is prevent you granting public access if you disable it

      It does not mean that the bucket is public. It just means that you can grant public access to this bucket

      So for this demonstration, we're going to go ahead and untick this option

      Now if you do untick this you'll need to scroll down and check this box just to acknowledge that you understand

      Exactly what you're doing

      So this is a safety feature of s3 that if you're going to remove this fail-safe check then you need to accept responsibility

      It means that if you do mistakenly grant public access to the bucket then potentially information can be exposed.

      Now I'm not going to explain any of these other options because I cover all of them in the S3 section of the course.

      So I'm going to skip past bucket versioning and tags,

      default encryption, and I'm not going to be covering any of these advanced settings.

      Instead, let's just go ahead and click on create bucket.

      At this point you might get an error that a bucket with the same name already exists, and that's fine.

      Remember S3 bucket names need to be globally unique

      And there's obviously a lot of koala campaigns happening in the wild. If you do get the error

      Then just feel free to add extra digits of random to the bucket name

      Then scroll all the way down to the bottom and create the bucket. Once the bucket's created

      You'll see it in the list of buckets

      So there's a column for the name a column for the region

      So you'll be able to see which region this bucket is in

      It will give you an overview of the access that this bucket has so because we unchecked the block public access

      Then it informs us that objects can be public again

      Just to stress this does not mean they are public because s3 buckets are private by default

      This is just telling us that they can be public

      Lastly, we also have the creation date which tells us when the bucket was created.

      So now let's just go ahead and click on the bucket to move inside so we can see additional information.

      Now one thing that I do want to draw your attention to is the Amazon resource name or ARN for this bucket.

      All resources in AWS have a unique identifier, the ARN or Amazon resource name.

      So this is the ARN for the bucket that we've just created

      ARNs have a consistent format. They start with ARN for Amazon resource name

      Then they have the partition for most

      AWS resources in most regions this will always say AWS

      Then you have the service name in this case S3

      Then you have some other values which I'll be talking about later in the course and you can omit those with certain

      services by just putting double colons. These for example might be the region or the account number.

      Now for services where resources are not globally unique

      then obviously you need to specify the region and the account number in order for this name to be globally unique.

      But because S3 buckets by default have to be globally unique

      then we don't have to specify in the ARN either

      the region or the account number. As long as we have the S3 service and the bucket name, we know that this uniquely

      references a resource and that's the key thing about ARNs. ARNs

      uniquely reference one resource within

      AWS. You always know if you have one ARN that it references one particular resource within AWS.

      Now you can use wildcards to reference multiple resources,

      but as a basis it has to reference at least one.

      Now let's just click on the objects tab here

      and this will give us an overview of all of the objects

      which are in this bucket.

      You have a number of tabs here that you can step through.

      We've got a properties tab where you can enable

      bucket versioning, tags, encryption, logging,

      CloudTrail data events, event notifications,

      transfer acceleration, object lock, request to pays,

      and static website hosting.

      and we'll be talking about all of those features in detail

      within the S3 section of the course.

      We'll also be covering permissions in that section

      because you can be very granular

      with the permissions of S3 buckets.

      You can see some metrics about the bucket.

      So this uses CloudWatch,

      which we'll be talking about in detail

      elsewhere in the course.

      You're also able to access management functionality.

      Again, we'll be talking about all of this

      later in the course.

      And then finally you're able to create access points.

      Now access points are some advanced functionality and so we'll be covering this later in the course.

      For now I just want you to get some experience of uploading some objects and interacting with them.

      Now there's a link attached to this lesson which you'll need to go ahead and click and that will download a zip file.

      Go ahead and extract that zip file and it will create a folder.

      And then once you've extracted that into a folder we're good to continue.

      Now the easiest way at this point to upload some objects is to make sure that you've got the objects tab selected

      and then click on upload.

      Now you're able to upload both files and folders to this S3 bucket.

      So let's start off by uploading some files.

      So click on add files.

      Now at this point locate and go inside the folder that you extracted a few moments ago

      and you'll see that there are three image files.

      We've got koala_nom1.jpg, koala_nom2.jpg and koala_zzz.jpg

      Now go ahead and select all three of these JPEG files and click on open.

      You'll see that you have three files in total queued for upload

      and you'll be provided with an estimate of the amount of space that these files will consume.

      Now scrolling down you're told the destination where you'll be uploading these objects to

      So this is the S3 bucket that we've created and this will be different for you.

      This will be your bucket name.

      Now we haven't enabled versioning on this bucket.

      This is a feature which I'll be covering in the S3 section of the course, but

      because we don't have versioning enabled, it means that if we do upload files with

      the same name, then potentially we're going to overwrite other objects in that bucket.

      So we have to accept the risk because we don't have versioning enabled.

      We could overwrite objects if we re-upload ones with the same name.

      In this case that's fine because we're not uploading anything important

      and regardless this bucket is empty so we can't overwrite anything.

      You have the option of enabling versioning where you can just acknowledge the risk.

      Then we can scroll down further still. We need to pick the storage class for the objects.

      This defaults to standard and I haven't covered storage classes in the course yet.

      I'll be doing that within the S3 section, so we're going to accept the default

      and then we're going to skip past all of these options.

      I'll be covering these later in the course and just go ahead and click on Upload.

      And this will upload all of those three objects to the S3 bucket.

      You'll be told whether the upload has been successful or whether it's failed.

      In our case, it's succeeded.

      So we can go ahead and click on Close to close down this dialogue.

      Now when we scroll down, we'll see an overview of the objects within this bucket.

      In our case we only have the three, Koala Nom1, Koala Nom2 and KoalaZZZ.jpg

      We can also create folders within S3. Now of course because S3 is a flat structure

      This isn't actually creating a folder. It's just creating a file which emulates a folder

      So if we create a folder and let's call this folder archive and then click on create folder

      It's not actually creating a folder called archive what it's doing is creating an object with this name

      so archive forward slash

      Now if we click on this archive folder and go inside it we can upload objects into this folder

      So let's go ahead and do that click on upload

      Go to add files and then just pick one of these files. Let's go with koala

      zz.jpg so select that one and click on open and just click on upload. Now what

      we've done is we've uploaded an object into what we see as a folder in this s3

      bucket. If we click on close what this has actually done is it's created an

      object which is called archive/koalazz.jpg. S3 doesn't really have

      folders. Folders are emulated using prefixes and that's important to know as

      you move through the course. Now if we click this at the top to go back to the

      main bucket and we're going to go ahead and open one of these objects. So let's

      pick one of these objects let's use koala_nom1.jpg. This opens an overview

      screen for this particular object and where we see this object URL just go

      ahead and right-click and open that in a new tab. When you open that in a new tab

      you'll be presented with an access denied error. The reason for that is

      you're trying to access this object with no authentication. You're accessing the

      object as an unauthenticated user and as I mentioned earlier all S3 objects and

      all S3 buckets are private by default and that's why you get this access

      denied you won't be able to access this object without authenticating to AWS and

      using that identity to access an object. That's of course unless you grant public

      access to this object which we haven't done and we won't be doing in this

      lesson. So if we close down that tab and instead click on open you might have to

      bypass a pop-up blocker but this time it will open this object and that's because

      we're including authentication in the URL at the top here. So when you click on

      the open button it's opening the object as you it's not opening it as an

      unauthenticated identity so that's important because you have access to

      this bucket you can open the objects using this open button. The same is true

      for the other objects so if we go back to the bucket let's pick koala nom2

      Then click on the open button and again we'll see a koala having some food.

      Go back to the bucket and then let's try koala ZZZ.

      So click on the object, click on open again and now we can see a koala having a well deserved

      rest after his lunch.

      Now that's everything I wanted to cover in this demo lesson.

      It's just been a really high level introduction into how to interact with S3 using the console

      UI.

      be covering S3 in detail later in the course, I just wanted this demo lesson to

      be a very brief introduction. Now what we need to do before we finish this demo

      lesson is to go back to the main S3 console and we need to tidy up by

      deleting this bucket. So deleting buckets within S3 is a two-step process. First we

      need to empty the bucket. So go ahead and select the bucket and click on empty.

      You'll need to either type or copy and paste permanently delete into the box

      and then click on empty and that will remove any objects within the bucket.

      Assuming that's successful go ahead and click on exit and with the bucket still

      selected click on delete and then once you've clicked on delete you'll need to

      copy and paste or type the name of the bucket and finally click on delete

      bucket to confirm that deletion process and that will delete the bucket and your

      account will be back in the same state as it was at the start of this demo

      lesson. Now at this point I hope this has been useful it's just been a really

      basic introduction to S3 and don't worry you'll be getting plenty more theory and

      practical exposure to the product in the S3 section of the course. For now just go

      ahead and complete this video and when you're ready I look forward to you

      joining me in the next.

    1. Gali WeinsteinPhD. Foundations (history, philosophy) of physics. · 1y · Here are the traits that might have prevented me from being accepeted to jobs for many years and also might have got me fired:Being a loner. I have the ability to be social but not too much. I have difficulty attending social meetings but I try to be friendly with people.Having a strong will, being rebellious, determined, and independent. I dropped out of a bona fide high school and I always drop out of bona fide schools. You need to attend bona fide schools to get jobs in the academy (I have a Ph.D). Throughout my studies at university, I had not attended boring classes. I am a rule breaker.Being autodidactic: teaching myself physics and mathematics, and being able to teach myself all sorts of things. People don't accept this.Having too much emotional empathy and being sensitive to people. I am sensitive to other people and have an extremely high sense of justice and fairness. I am intense in helping other people but they are not so intense in helping me…. But I burn bridges with people because I do not fit in and they reject “Good doctors” (i. e. Dr. Shaun Murphy). Dr. Murphy says in the “Good Doctor”: "There is a long and dusty trail littered with people who have underestimated me". Precisely.I have a sense of humor which people find hard to understand, and I am child-like.I show attention to detail and I also tend to correct mistakes I find in other people's papers.I have difficulty understanding that I have embarrassed people and they have difficulty understanding me. I am saying what I think, and am unable to hide my feelings. I am overly direct and frank. Dr. Murphy was asked: "Do you always talk like that? Just say whatever you're thinking when you're thinking it?" And he answered: "Yes. It's good to be honest". "I have autism, it's part of who I am". So, people are never going to crowd in our corner. But people tell Dr. Murphy that he is smart, brutally honest, has no regard to social convention, and has a problem being a leader; that he is facetious, he can't understand what anyone means. He can't express himself like an adult. His apparently robotic voice is not a charming affectation. Yeah, but I don’t have a robotic voice.I constantly feel I am taken advantage of because I trust people, and they insult me.I have a very good mastery of Einstein’s classical general relativity, and I am engaging in my special interest which is Einstein and relativity, and people are jealous. But this is the kind of obsessive behavior that is common for autism. And part of being obssesive is being highly imaginative and inventive. That is, I am highly creative and have a burst of ideas in my head. I am able to concentrate on my work for a long period of time without eating and drinking.I have skills that are much higher than those my job requires. I am not a professor. I have a lot of setbacks and very few successes.My academic papers are often times not conventional. I am not writing papers with a group of scholars. I am not part of a group of scholars. It’s not because I don’t want to be part of a group of scholars.I misinterpret the implied meaning of things and take at face value what people tell me, and I overreact and over-analyze situations. I don’t understand the unwritten social rules and ask embarrassing questions. I am socially naïve, and gullible, and I believe what people tell me. Consequently, I am taken advantage of and have no understanding of how to get on with important people. For many years, I was vulnerable to bullying. My name has been conspicuously absent from the list of speakers of conferences and seminars. I felt that people picked on me and I felt great indignation. I have been a recluse for many years, spending a lot of time alone in my room. I thought that I would not get any academic position because it involves social and political skills. I don't know how to grovel and don’t know whether one actually needs to grovel to bigwigs.I prefer wearing something casual over an elegant outfit.I know I am right but I don’t know how I know this. I have this intuition, six sense…. I also have a very good visual memory.I try to mask my real personality traits.
      • NOT: 4, 8, 15
      • ?: 7
      • : 1, 3, 5, 6, 13

      • = *
    1. Reviewer #1 (Public Review):

      Summary:

      The question of whether eyespots mimic eyes has certainly been around for a very long time and led to a good deal of debate and contention. This isn't purely an issue of how eyespots work either, but more widely an example of the potential pitfalls of adopting 'just-so-stories' in biology before conducting the appropriate experiments. Recent years have seen a range of studies testing eye mimicry, often purporting to find evidence for or against it, and not always entirely objectively. Thus, the current study is very welcome, rigorously analysing the findings across a suite of papers based on evidence/effect sizes in a meta-analysis.

      Strengths:

      The work is very well conducted, robust, objective, and makes a range of valuable contributions and conclusions, with an extensive use of literature for the research. I have no issues with the analysis undertaken, just some minor comments on the manuscript. The results and conclusions are compelling. It's probably fair to say that the topic needs more experiments to really reach firm conclusions but the authors do a good job of acknowledging this and highlighting where that future work would be best placed.

      Weaknesses:

      There are few weaknesses in this work, just some minor amendments to the text for clarity and information.

    1. Reviewer #2 (Public Review):

      Summary:

      This paper attempts to examine how rare, extreme events impact decision-making in rats. The paper used an extensive behavioural study with rats to evaluate how the probability and magnitude of outcomes impact preference. The paper, however, provides limited evidence for the conclusions because the design did not allow for the isolation of the rare, extreme events in choice. There are many confounding factors, including the outcome variance and presence of less-rare, and less-extreme outcomes in the same conditions.

      Strengths:

      (1) The major strength of the paper is the significant volume of behavioural data with a reasonable sample size of 20 rats.

      (2) The paper attempts to examine losses with rats (a notoriously tricky problem with non-human animals) by substituting time-outs as a proxy for losses. This allows for mixed gambles that have both gain and loss possible outcomes.

      (3) The paper integrates both a behavioural and a modelling approach to get at the factors that drive decision-making.

      (4) The paper takes seriously the question of what it means for an event to be rare, pushing to less frequent outcomes than usually used with non-human animals.

      Weaknesses:

      (1) The primary issue with this work is that the primary experimental manipulation fails to isolate the rare, extreme events in choice. As I understand the task, in all the conditions with a rare extreme event (e.g., 80 pellets with probability epsilon), there is also a less-rare, less-extreme event (e.g., 12 pellets with probability 5). In addition, the variance differs between the two conditions. So, any impact attributable to the rare, extreme event could be due to the less rare event or due difference in the variance. The design does not support the conclusions. Finally, by deliberately confounding rarity and extremity, the design does not allow for assessing the impact of either aspect.

      (2) The RL-modelling work also fails to show a specific impact of the rare extreme event. As best as I can understand Eq 2, the model provides a free parameter that adds a bonus to the value of either the two options with high-variance gains (A and V in the paper) or to the two options with high-variance losses (F and V in the paper). This parameter only depends on whether this option could have possibly yielded the rare, extreme outcome (i.e., based on the generative probability) and was not connected to its actual appearance. That makes it a free parameter that just bumps up (or down) the probability of selecting a pair of options. In the case of the "black swan" or high-variance loss conditions, this seems very much like a loss aversion parameter, but an additive one instead of a multiplicative one.

      (3) The paper presented the methods and results with lots of neologisms and fairly obscure jargon (e.g., fragility, total REE sensitivity). That made it very hard to decipher exactly what was done and what was found. For example, on p. 4, the use of concave and convex was very hard to decipher; the text even has to repeat itself 3 times (i.e., "to repeat" and "in other words") and is still not clear. It would be much clearer (and probably accurate) to say that the options varied along the variance dimension, separately for gains and losses. Option A was low-variance gains and losses. Option B was low-variance losses and high-variance gains. Option C was high-variance losses and low-variance gains, and Option D was high-variance losses and gains. That tells much more clearly what the animals experienced without the reader having to master a set of new terminologies around fragility and robustness, which brings a set of theoretical assumptions unnecessarily into the description of the experimental design. In terms of results, "Black Swan" avoidance is more simply known as risk aversion for losses.

      (4) Were the probabilities shuffled or truly random (seem to be fixed sequences, so neither)? What were the experienced probabilities? Given the fixed sequences, these experienced ("ex-post") probabilities, could differ tremendously from the scheduled ("ex ante") probabilities. It's quite possible that an animal never experienced the rare, extreme event for a specific option. It's even possible (if they only picked it on the 10th/60th choices by chance), that they only ever experienced that rare extreme event. This cannot be known given the information provided. The Supplemental info on p.55 only gives gross overall numbers but does not indicate what the rats experienced for each choice/option-which is what matters here. A simple table that indicates for each of the 4 options, how often they were selected, and how often the animals experienced each of the 6-8 possible outcome would make it much clearer how closely the experience matched the planned outcomes. In addition, by restricting the rare outcome to either the 10th or 60th activations in a session, these are not random. Did the animals learn this association?

      (5) The choice data are only presented in an overprocessed fashion with a sum and a difference (in both figures and tables). The basic datum (probability/frequency of selecting each of the 4 options) is not provided directly, even if it can theoretically be inferred from the sum and the difference. To understand what the rats actually do, we first need to see how often they select each option, without these transformations.

      (6) There is insufficient detail provided on the inferential statistical tests (e.g., no degrees of freedom or effect sizes), and only limited information on exactly what tests were run and how (bootstrapping, but little detail). Without code or data (only summary information is provided in the supplement), this is difficult to evaluate. In addition, the studies seem not to be pre-registered in any way, leaving many researchers with degrees of freedom. Were any alternative analysis pipelines attempted? Similarly, there were many sub-groupings of the animals, and then comparisons between them - were these post-hoc?

      (7) On p. 17, there is an attempt to look at the impact of a rare, extreme event by plotting a measure of preference for the 10 trials before/after the rare, extreme event. In the human literature, the main impact of experiencing a rare, extreme event is what is known as the wavy recency effect (See Plonsky et al. 2015 in Psych Review for example). What this means is that there tends to be some immediate negative recency (e.g., avoiding a rare gain) followed by positive recency (e.g., chasing the rare gain). Using a 10-trial window would thus obscure any impact of this rare, extreme event. An analysis that looks at a time course trial-by-trial could reveal any impact.

      (8) As I understood the method (p. 31), the assignment of options to physical locations was not random or counterbalanced, but deliberately biased to have one of the options in the preferred location. This would seem to create a bias towards a particular option and a bias away from the other options, which confounds the preference data in subsequent analyses.

      (9) Are delays really losses? This is a big assumption. Magnitude and delay are different aspects of experience, which are not necessarily commensurable and can be manipulated independently. And, for the model, how were these delays transformed into outcomes for the model? Eq 1 skips over that. Is there an assumption of linearity? In addition, I was not wholly clear if the delays meant fewer trials in a session or if the delays merely extended the session and meant longer delays until the next choice period.

      (10) The paper does not sufficiently accurately represent the existing literature on human risky decision-making (with and without rare events). Here are a few examples of misrepresented and/or missing literature:<br /> -Most studies on decision-making do not only rely on p > 10% (as per p. 2). Maybe that is true with animals, but not a fair statement generally. Some do, and some don't. There is substantial literature looking at rarer events in both descriptions (most famously with Kahneman & Tversky's work), but also in experience (which is alluded to in reference 19). That reference is not only about the situation when choices are not repeated (e.g. the sampling paradigm), but also partial feedback and full-feedback situations.

      The literature on learning from rewarding experiences in humans is obliquely referenced but not really incorporated. In short, there are two main findings - firstly people underweight rare events in experience; second, people overweight extreme outcomes in experience (both contrary to description). Some related papers are cited, but their content is not used or incorporated into the logic of the manuscript.

      One recent study systematically examined rarity and extremity in human risky decision-making, which seems very relevant here: Mason et al. (2024). Rare and extreme outcomes in risky choice. Psychonomic Bulletin & Review, 31, 1301-1308.

      There is a fair bit of research on the human perception of the risk of rare events (including from experience) and important events like climate. One notable paper is Newell et al (2015) in Nature Climate Change.

    1. Reviewer #3 (Public Review):

      Summary:

      The authors performed wide-field and 2-photon imaging in vivo in awake head-fixed mice, to compare receptive fields and tonotopic organization in thalamocortical recipient (TR) neurons vs corticothalamic (CT) neurons of mouse auditory cortex. TR neurons were found in all cortical layers while CT neurons were restricted to layer 6. The TR neurons at nominal depths of 200-400 microns have a remarkable degree of tonotopy (as good if not better than tonotopic maps reported by multiunit recordings). In contrast, CT neurons were very heterogenous in terms of their best frequency (BF), even when focusing on the low vs high-frequency regions of the primary auditory cortex. CT neurons also had wider tuning.

      Strengths:

      This is a thorough examination using modern methods, helping to resolve a question in the field with projection-specific mapping.

      Weaknesses:

      There are some limitations due to the methods, and it's unclear what the importance of these responses are outside of behavioral context or measured at single timepoints given the plasticity, context-dependence, and receptive field 'drift' that can occur in the cortex.

      (1) Probably the biggest conceptual difficulty I have with the paper is comparing these results to past studies mapping auditory cortex topography, mainly due to differences in methods. Conventionally, the tonotopic organization is observed for characteristic frequency maps (not best frequency maps), as tuning precision degrades and the best frequency can shift as sound intensity increases. The authors used six attenuation levels (30-80 dB SPL) and reported that the background noise of the 2-photon scope is <30 dB SPL, which seems very quiet. The authors should at least describe the sound-proofing they used to get the noise level that low, and some sense of noise across the 2-40 kHz frequency range would be nice as a supplementary figure. It also remains unclear just what the 2-photon dF/F response represents in terms of spikes. Classic mapping using single-unit or multi-unit electrodes might be sensitive to single spikes (as might be emitted at characteristic frequency), but this might not be as obvious for Ca2+ imaging. This isn't a concern for the internal comparison here between TR and CT cells as conditions are similar, but is a concern for relating the tonotopy or lack thereof reported here to other studies.

      (2) It seems a bit peculiar that while 2721 CT neurons (N=10 mice) were imaged, less than half as many TR cells were imaged (n=1041 cells from N=5 mice). I would have expected there to be many more TR neurons even mouse for mouse (normalizing by number of neurons per mouse), but perhaps the authors were just interested in a comparison data set and not being as thorough or complete with the TR imaging?

      (3) The authors' definitions of neuronal response type in the methods need more quantitative detail. The authors state: ""Irregular" neurons exhibited spontaneous activity with highly variable responses to sound stimulation. "Tuned" neurons were responsive neurons that demonstrated significant selectivity for certain stimuli. "Silent" neurons were defined as those that remained completely inactive during our recording period (> 30 min). For tuned neurons, the best frequency (BF) was defined as the sound frequency associated with the highest response averaged across all sound levels.". The authors need to define what their thresholds are for 'highly variable', 'significant', and 'completely inactive'. Is best frequency the most significant response, the global max (even if another stimulus evokes a very close amplitude response), etc.

    1. Reviewer #1 (Public Review):

      Summary:

      In this study, the authors identified and described the transcriptional trajectories leading to CMs during early mouse development, and characterized the epigenetic landscapes that underlie early mesodermal lineage specification.

      The authors identified two transcriptomic trajectories from a mesodermal population to cardiomyocytes, the MJH and PSH trajectories. These trajectories are relevant to the current model for the First Heart Field (FHF) and the Second Heart Field (SHF) differentiation. Then, the authors characterized both gene expression and enhancer activity of the MJH and PSH trajectories, using a multiomics analysis. They highlighted the role of Gata4, Hand1, Foxf1, and Tead4 in the specification of the MJH trajectory. Finally, they performed a focused analysis of the role of Hand1 and Foxf1 in the MJH trajectory, showing their mutual regulation and their requirement for cardiac lineage specification.

      Strengths:

      The authors performed an extensive transcriptional and epigenetic analysis of early cardiac lineage specification and differentiation which will be of interest to investigators in the field of cardiac development and congenital heart disease. The authors considered the impact of the loss of Hand1 and Foxf1 in-vitro and Hand1 in-vivo.

      Weaknesses:

      The authors used previously published scRNA-seq data to generate two described transcriptomic trajectories.

      (1) Details of the re-analysis step should be added, including a careful characterization of the different clusters and maker genes, more details on the WOT analysis, and details on the time stamp distribution along the different pseudotimes. These details would be important to allow readers to gain confidence that the two major trajectories identified are realistic interpretations of the input data.

      The authors have also renamed the cardiac trajectories/lineages, departing from the convention applied in hundreds of papers, making the interpretation of their results challenging.

      (2) The concept of "reverse reasoning" applied to the Waddington-OT package for directional mass transfer is not adequately explained. While the authors correctly acknowledged Waddington-OT's ability to model cell transitions from ancestors to descendants (using optimal transport theory), the justification for using a "reverse reasoning" approach is missing. Clarifying the rationale behind this strategy would be beneficial.

      (3) As the authors used the EEM cell cluster as a starting point to build the MJH trajectory, it's unclear whether this trajectory truly represents the cardiac differentiation trajectory of the FHF progenitors:<br /> - This strategy infers that the FHF progenitors are mixed in the same cluster as the extra-embryonic mesoderm, but no specific characterization of potential different cell populations included in this cluster was performed to confirm this.

      - The authors identified the EEM cluster as a Juxta-cardiac field, without showing the expression of the principal marker Mab21l2 per cluster and/or on UMAPs.

      - As the FHF progenitors arise earlier than the Juxta-cardiac field cells, it must be possible to identify an early FHF progenitor population (Nkx2-5+; Mab21l2-) using the time stamp. It would be more accurate to use this FHF cluster as a starting point than the EEM cluster to infer the FHF cardiac differentiation trajectory.

      These concerns call into question the overall veracity of the trajectory analysis, and in fact, the discrepancies with prior published heart field trajectories are noted but the authors fail to validate their new interpretation. Because their trajectories are followed for the remainder of the paper, many of the interpretations and claims in the paper may be misleading. For example, these trajectories are used subsequently for annotation of the multiomic data, but any errors in the initial trajectories could result in errors in multiomic annotation, etc, etc.

      (4) As mentioned in the discussion, the authors identified the MJH and PSH trajectories as non-overlapping. But, the authors did not discuss major previously published data showing that both FHF and SHF arise from a common transcriptomic progenitor state in the primitive streak (DOI: 10.1126/science.aao4174; DOI: 10.1007/s11886-022-01681-w). The authors should consider and discuss the specifics of why they obtained two completely separate trajectories from the beginning, how these observations conflict with prior published work, and what efforts they have made at validation.

      (5) Figures 1D and E are confusing, as it's unclear why the authors selected only cells at E7.0. Also, panels 1D 'Trajectory' and 'Pseudotime' suggest that the CM trajectory moves from the PSH cells to the MJH. This result is confusing, and the authors should explain this observation.

      (6) Regarding the PSH trajectory, it's unclear how the authors can obtain a full cardiac differentiation trajectory from the SHF progenitors as the SHF-derived cardiomyocytes are just starting to invade the heart tube at E8.5 (DOI: 10.7554/eLife.30668).

      The above notes some of the discrepancies between the author's trajectory analysis and the historical cardiac development literature. Overall, the discrepancies between the author's trajectory analysis and the historical cardiac development literature are glossed over and not adequately validated.

      (7) The authors mention analyzing "activated/inhibited genes" from Peng et al. 2019 but didn't specify when Peng's data was collected. Is it temporally relevant to the current study? How can "later stage" pathway enrichment be interpreted in the context of early-stage gene expression?

      (8) Motif enrichment: cluster-specific DAEs were analyzed for motifs, but the authors list specific TFs rather than TF families, which is all that motif enrichment can provide. The authors should either list TF families or state clearly that the specific TFs they list were not validated beyond motifs.

      (9) The core regulatory network is purely predictive. The authors again should refrain from language implying that the TFs in the CRN have any validated role.

      Regarding the in vivo analysis of Hand1 CKO embryos, Figures 6 and 7:

      (10) How can the authors explain the presence of a heart tube in the E9.5 Hand1 CKO embryos (Figure 6B) if, following the authors' model, the FHF/Juxta-cardiac field trajectory is disrupted by Hand1 CKO? A more detailed analysis of the cardiac phenotype of Hand1 CKO embryos would help to assess this question.

      (11) The cell proportion differences observed between Ctrl and Hand1 CKO in Figure 6D need to be replicated and an appropriate statistical analysis must be performed to definitely conclude the impact of Hand1 CKO on cell proportions.

      (12) The in-vitro cell differentiations are unlikely to recapitulate the complexity of the heart fields in-vivo, but they are analyzed and interpreted as if they do.

      (13) The schematic summary of Figure 7F is confusing and should be adjusted based on the following considerations:<br /> (a) the 'Wild-type' side presents 3 main trajectories (SHF, Early HT and JCF), but uses a 2-color code and the authors described only two trajectories everywhere else in the article (aka MJH and PSH). It's unclear how the SHF trajectory (blue line) can contribute to the Early HT, when the Early HT is supposed to be FHF-associated only (DOI: 10.7554/eLife.30668). As mentioned previously in Major comment 3., this model suggests a distinction between FHF and JCF trajectories, which is not investigated in the article.<br /> (b) the color code suggests that the MJH (FHF-related) trajectory will give rise to the right ventricle and outflow tract (green line), which is contrary to current knowledge.

      Minor comments:

      (1) How genes were selected to generate Figure 1F? Is this a list of top differentially expressed genes over each pseudotime and/or between pseudotimes?

      (2) Regarding Figure 1G, it's unclear how inhibited signaling can have an increased expression of underlying genes over pseudotimes. Can the authors give more details about this analysis and results?

      (3) How do the authors explain the visible Hand1 expression in Hand1 CKO in Figure S7C 'EEM markers'? Is this an expected expression in terms of RNA which is not converted into proteins?

      (4) The authors do not address the potential presence of doublets (merged cells) within their newly generated dataset. While they mention using "SCTransform" for normalization and artifact removal, it's unclear if doublet removal was explicitly performed.

    1. Note: This response was posted by the corresponding author to Review Commons. The content has not been altered except for formatting.

      Learn more at Review Commons


      Reply to the reviewers

      Manuscript number: RC-2024-02394

      Corresponding author(s): Altman, Brian J

      1. General Statements [optional]

      We thank all three Reviewers for their insightful and helpful feedback and suggestions. We strongly believe that addressing these comments has now resulted in a much-improved manuscript. We appreciate that the Reviewers found the manuscript "interesting" with "valuable insights and... obvious novelty", "an important study that is well-done", and "an important understanding of the crosstalk between cancer cells and immune cells as well as the understanding of how the TME disrupts circadian rhythms". All three Reviewers requested a significant revision, which we provide here. We carefully and completely responded to each Reviewer question or suggestion, in most cases with new experiments and text, and in a very few cases with changes or additions to the Discussion section. This includes new data in seven of the original Figures and Supplementary Figures, and one new main Figure and three new Supplementary Figures. Highlights of these new data include testing the role of low pH in cancer cell supernatant on macrophage rhythms, and analysis of single-cell RNA-sequencing data for heterogeneity in macrophage circadian gene expression. Additional experiments were also performed that were not included in the manuscript, and these data are presented in this Response. A detailed point-by-point response to each comment is included below with excerpts of the data and updated text for the reviewers. Please note that the PDF version of this Response includes images of the new Figures inserted in to the manuscript.

      2. Point-by-point description of the revisions

      __Reviewer #1 __

      Evidence, reproducibility and clarity

      The manuscript by Knudsen-Clark et al. investigates the novel topic of circadian rhythms in macrophages and their role in tumorigenesis. The authors explore how circadian rhythms of macrophages may be influenced by the tumor microenvironment (TME). They utilize a system of bone marrow-derived macrophages obtained from transgenic mice carrying PER2-Luciferase (PER2-Luc), a trackable marker of rhythmic activity. The study evaluates how conditions associated with the TME, such as polarizing stimuli (to M1 or M2 subtype), acidic pH, and elevated lactate, can each alter circadian rhythms in macrophages. The authors employ several approaches to explore macrophage functions in cancer-related settings. While the manuscript presents interesting findings and may be the first to demonstrate that tumor stimuli alter circadian rhythms in macrophages and impact tumor growth, it lacks a clear conclusion regarding the role of altered circadian rhythms in suppressing tumor growth. Several discrepancies need to be addressed before publication, therefore, the manuscript requires revision before publication, addressing the following comments:

      We thank Reviewer #1 for the comments regarding the quality of our work and are pleased that the Reviewer finds that this manuscript "presents interesting findings and may be the first to demonstrate that tumor stimuli alter circadian rhythms in macrophages and impact tumor growth". We have addressed all comments and critiques from Reviewer #1 below. To summarize, we added new data on how different macrophage polarization states affect media pH (Supplementary Figure 4), further characterized gene expression in our distinct macrophage populations (Supplementary Figure 1), provided clarity in the data and text on the universal nature of Clock Correlation Distance (CCD) across macrophage populations (Figure 6), included human tumor-associated macrophage (TAM) data for CCD (Figure 7) analyzed single-cell RNA-sequencing data of TAMs to demonstrate heterogeneity in circadian gene expression (Figure 9), and used tumor-conditioned media to show that low pH still affects macrophage rhythms in this context *Supplementary Figure 5". Thanks to the helpful suggestions of the Reviewer, we also made numerous clarifications and fixed a critical referencing error that the Reviewer identified.

      Major comments: 1. It is well known that pro-inflammatory macrophages primarily rely on glycolysis during inflammation, exhibiting dysregulated tricarboxylic acid (TCA) cycle activity. These pro-inflammatory macrophages are commonly referred to as 'M1' or pro-inflammatory, as noted in the manuscript. In contrast, M2 macrophages, or pro-resolution macrophages, are highly dependent on active mitochondrial respiration and oxidative phosphorylation (OXPHOS). Given that M1 macrophages favor glycolysis, they create an acidic environment due to elevated lactate levels and other acidifying metabolites. However, the study does not address this effect. The authors' hypothesis revolves around the acidic environment created by glycolytic tumors, yet they overlook the self-induced acidification of media when culturing M1 macrophages. This raises the question of how the authors explain the reduced circadian rhythms observed in pro-inflammatory macrophages in their study, while low pH and higher lactate levels enhance the amplitude of circadian rhythms. I would encourage the authors to incorporate the glycolytic activity of pro-inflammatory macrophages into their experimental setup. Otherwise the data look contradictory and misleading in some extent.

      We appreciate the important point Reviewer #1 made that macrophages polarized toward a pro-inflammatory phenotype such as those stimulated with IFNγ and LPS (M1 macrophages) prioritize metabolic pathways that enhance glycolytic flux, resulting in increased release of protons and lactate as waste products from the glycolysis pathway. In this way, polarization of macrophages toward the pro-inflammatory phenotype can lead to acidification of the media, which may influence our observations given that we are studying the effect of extracellular pH on rhythms in macrophages. To address this point, we have performed additional experiments in which we measured pH of the media to capture changes in media pH that occur during the time in which we observe changes in rhythms of pro-inflammatory macrophages.

      In line with the documented enhanced glycolytic activity of pro-inflammatory macrophages, the media of pro-inflammatory macrophages is acidified over time, in contrast to media of unstimulated or pro-resolution macrophages. Notably, while pH decreased over time in the pro-inflammatory group, the pH differential between the pH7.4, pH6.8, and pH6.5 sample groups was maintained over the period in which we observe and measure changes in circadian rhythms of pro-inflammatory macrophages. Additionally, media that began at pH 7.4 was acidified only to pH 7 by day 2, above the acidic pH of 6.8 or 6.5. As a result, there remained a difference in pH between the two groups (pH 7.4 and pH 6.5) out to 2 days consistent with the changes in rhythms that we observe between these two groups. This indicates that the difference in circadian rhythms observed in pro-inflammatory macrophages cultured at pH 7.4 compared to pH 6.5 were indeed due to the difference in extracellular pH between the two conditions. We have incorporated these data, shown below, into Supplementary Figure 4 and added the following discussion of these data to the Results section:

      "In line with their documented enhanced glycolytic capacity, pro-inflammatory macrophages acidified the media over time (Supplementary Figure 4C). Notably, while pH of the media the pro-inflammatory macrophages were cultured in decreased over time pH, the pH differential between the pH 7.4, pH 6.8, and pH 6.5 samples groups of pro-inflammatory macrophages was maintained out to 2 days, consistent with the changes in rhythms that we observe and measure between these groups."

      The article examines the role of circadian rhythms in tumor-associated macrophages, yet it lacks sufficient compelling data to support this assertion. Two figures, Figure 7 and Figure 9, are presented in relation to cancer. In Figure 7, gene expression analysis of Arg1 (an M2 marker) and Crem (a potential circadian clock gene) is conducted in wild-type macrophages, BMAL1-knockout macrophages with dysregulated circadian rhythms, and using publicly available data on tumor-associated macrophages from a study referenced as 83. However, it is noted that this referenced study is actually a review article by Geeraerts et al. (2017) titled "Macrophage Metabolism as Therapeutic Target for Cancer, Atherosclerosis, and Obesity" published in Frontiers in Immunology. This raises concerns about the reliability of the results. Furthermore, comparing peritoneal macrophages from healthy mice with macrophages isolated from lung tumors is deemed inaccurate. It is suggested that lung macrophages from healthy mice and those from mice with lung tumors should be isolated separately for a more appropriate comparison. Consequently, Figure 7B is further questioned regarding how the authors could compare genes from the circadian rhythm pathway between these non-identical groups. As a result, the conclusion drawn from these data, suggesting that tumor-associated macrophages exhibit a gene expression pattern similar to BMAL1-KO macrophages, is deemed incorrect, affecting the interpretation of the data presented in Figure 8.

      We thank Reviewer #1 for pointing out our error in the reference provided as the source of the TAM data used for CCD in Figure 7. While we took care to provide the GEO ID for the data set (GSE188549) in the Methods section, we mistakenly cited Geeraerts (2017) Front Immunol when we should have cited Geeraerts (2021) Cell Rep. We have corrected this citation error in the main text.

      We also appreciate Reviewer #1's concern that we are comparing circadian gene expression of peritoneal macrophages to tumor-associated macrophages derived from LLC tumors, which are grown ectopically in the flank for the experiment from which the data set was produced. To ensure an accurate comparison of gene expression, we downloaded the raw FASTQ files from each dataset and processed them in identical pipelines. Our main comparison between these cell types is Clock Correlation Distance (CCD), which compares the pattern of co-expression of circadian genes (Shilts et al PeerJ 2018). CCD was built from multiple mouse and human tissues to be a "universal" tool to compare circadian rhythms, and designed to compare between different tissues and cell types. Each sample is compared to a reference control built from these multiple tissues. To better convey this concept to readers to give confidence the suitability of CCD for comparing data sets across different tissues, we have added the reference control to Figure 7 (now Figure 6B), We have also expanded our analysis to include bone marrow-derived macrophages, to further demonstrate that the organization of clock gene co-expression is not specific to peritoneal macrophages; we have added this data to Figure 7 (now Figure 6C,D). Finally, we have included an abbreviated explanation of the points made above in the results section.

      Due to the universal nature of the CCD tool, we disagree with Reviewer #1's assertion that "the conclusion drawn from these data, suggesting that tumor-associated macrophages exhibit a gene expression pattern similar to BMAL1-KO macrophages, is deemed incorrect". Indeed, this finding mirrors findings in the original CCD paper, which showed that tumor tissues universally exhibit a disordered molecular clock as compared to normal tissue. Notably, the original CCD paper also compared across cell and tumor types.

      As an additional note to the review, we would like to clarify that nowhere in the manuscript do we propose that Crem is a potential circadian clock gene. We are clear throughout the manuscript that we are using Crem as a previously established biomarker for acidic pH-sensing in macrophages. Please see below for the modified Figure and text.

      "To understand the status of the circadian clock in TAMs, we performed clock correlation distance (CCD) analysis. This analysis has previously been used to assess functionality of the circadian clock in whole tumor and in normal tissue[102]. As the circadian clock is comprised of a series of transcription/translation feedback loops, gene expression is highly organized in a functional, intact clock, with core clock genes existing in levels relative to each other irrespective of the time of day. In a synchronized population of cells, this ordered relationship is maintained at the population level, which can be visualized in a heatmap. CCD is designed to compare circadian clock gene co-expression patterns between different tissues and cell types. To accomplish this, CCD was built using datasets from multiple different healthy tissues from mouse and human to be a universal tool to compare circadian rhythms. Each sample is compared to a reference control built from these multiple tissues (Figure 6B)[102]. To validate the use of this analysis for assessing circadian disorder in macrophages, we performed CCD analysis using publicly available RNA-sequencing data from bone marrow-derived macrophages and wild type peritoneal macrophages, as a healthy control for functional rhythms in a synchronized cell population, and BMAL1 KO peritoneal macrophages, as a positive control for circadian disorder[44]."

      And in the Discussion:

      "Interestingly, analysis of TAMs by clock correlation distance (CCD) presents evidence that rhythms are disordered in bulk TAMs compared to other macrophage populations (Figure 6). CCD is one of the most practical tools currently available to assess circadian rhythms due to its ability to assess rhythms independent of time of day and without the need for a circadian time series, which is often not available in publicly available data from mice and humans[102]."

      If the authors aim to draw a clear conclusion regarding the circadian rhythms of tumor-associated macrophages (TAMs), they may need to analyze single-sorted macrophages from tumors and corresponding healthy tissues. Such data are publicly available (of course not in #83)

      We agree with Reviewer #1 that while our interpretation of the data is that there may be heterogeneity in circadian rhythms of tumor-associated macrophages, we cannot prove this without assessing circadian rhythms at the single cell level. While single-cell RNA-sequencing data of freshly isolated tumor associated macrophages of sufficient read depth for circadian gene expression analysis has historically been unavailable, fortunately a dataset was released recently (May 2024) which we were able to use to address this point. We have analyzed publicly available single-cell RNAseq data of tumor-associated macrophages (GSE260641, Wang 2024 Cell) to determine whether there are differences in expression of circadian clock genes between different TAM populations. We have added these data as a new Figure 9. Please see the figure and updated text below.

      "Tumor-associated macrophages exhibit heterogeneity in circadian clock gene expression.

      __ Our findings suggested that heterogeneity of the circadian clock may lead to disorder in bulk macrophage populations, but did not reveal if specific gene expression changes exist in tumor-associated macrophages at the single-cell level. To determine whether heterogeneity exists within the expression of circadian clock genes of the tumor-associated macrophage population, we analyzed publicly available single-cell RNA sequencing data of macrophages isolated from B16-F10 tumors[107]. To capture the heterogeneity of macrophage subsets within the TAM population, we performed unbiased clustering (Figure 9A). We then performed differential gene expression to determine if circadian clock genes were differentially expressed within the TAM subpopulations. The circadian clock genes Bhlhe40 (DEC1), Bhlhe41 (DEC2), Nfil3 (E4BP4), Rora (RORα), Dbp (DBP), and Nr1d2 (REV-ERBβ) were significantly (adj.p We next sought to determine whether differences in circadian clock gene expression between TAM subpopulations were associated with exposure to acidic pH in the TME. To this end, we first assessed Crem expression in the TAM subpopulations that were identified by unbiased clustering. Crem expression was significantly higher in TAM clusters 4, 5, and 6 compared to TAM clusters 1-3 and 7-9 (Figure 9C). Clusters were subset based on Crem expression into Crem high (clusters 4-6) and Crem low (clusters 1-3, 7-9) (Figure 9D), and differential gene expression analysis was performed. The circadian clock genes Nfil3, Rora, Bhlhe40, and Cry1 (CRY1) were significantly (adj.p __And in the Discussion:

      "Supporting the notion that population-level disorder may exist in TAMs, we used scRNA-sequencing data and found evidence of heterogeneity between the expression of circadian clock genes in different TAM subpopulations (Figure 9A, B). Phenotypic heterogeneity of TAMs in various types of cancer has previously been shown[20, 21, 125, 126], and we have identified distinct TAM subpopulations by unbiased clustering (Figure 9A). Within those TAM subpopulations, we identified differential expression of circadian clock genes encoding transcription factors that bind to different consensus sequences: DEC1 and DEC2 bind to E-boxes, NFIL3 and DBP binds to D-boxes, and RORα and REV-ERBβ binds to retinoic acid-related orphan receptor elements (ROREs)[127, 128]. While little is known about regulation of macrophages by E-box and D-box elements beyond the circadian clock, aspects of macrophage function have been shown to be subject to transcriptional regulation through ROREs[129, 130]. Thus, we speculate that variations in these transcription factors may exert influence on expression of genes to drive diversity between TAM subpopulations. Differential expression of circadian clock genes between TAM subpopulations was also associated with Crem expression (Figure 9C-E), suggesting that exposure of TAMs to acidic pH within the TME can alter the circadian clock. However, there remained significant variation in expression of circadian clock genes within the Crem high and Crem low groups (Figure 9B), suggesting that acidic pH is not the only factor in the TME that can alter the circadian clock. Together, these data implicate the TME in driving heterogeneity in TAM circadian rhythms just as it drives heterogeneity in TAM phenotype.

      Interestingly, in contrast to our observations of circadian disorder in TAMs isolated from LLC tumors (Figure 6), rhythmicity in expression of circadian genes was observed in bulk TAMs isolated from B16 tumors[107]. This suggests that circadian rhythms of TAMs are maintained differently in different types of cancer. Notably, both of these observations were at the population level. Upon separation of the B16 TAM population into subsets by unbiased clustering of single-cell RNA sequencing data, we measured differences in expression of circadian clock genes between TAM subpopulations (Figure 9A,B). This suggests that even within a rhythmic TAM population, there is heterogeneity in the circadian clock of TAM subpopulations."

      Additionally, it is widely acknowledged that human and mouse macrophages exhibit distinct gene expression profiles, both in vitro and in vivo. While assuming that genes involved in circadian rhythms are conserved across species, the authors could consider extending their funding to include analyses of single-sorted macrophages from cancer patients, such as those with lung cancer or pancreatic ductal adenocarcinoma (PDAC). These experiments would provide relevant insights into TAM biology.

      We agree that with Reviewer #1 that ultimately, being able to relate findings in mice to humans is critical. It is important to assess if circadian disorder is observed in TAMs in human cancers as it is for LLC tumor-derived macrophages in mice. To address this point, we have performed CCD using a human data set (GSE116946; Garrido 2020 J Immunother Cancer) suitable for use with CCD (wherein macrophages were isolated from bulk tumor in humans, with a high enough samples size, and not cultured prior to sequencing). We have added these data as a new Figure 7, shown below. Please see the added data and updated text below.

      "We next assessed the status of the circadian clock in human TAMs from NSCLC patients. We performed CCD with publicly available RNA-seq data of tumor-adjacent macrophages and tumor-associated macrophages from NSCLC patients, using alveolar macrophages from healthy donors as a control[104, 105]. To assess the contribution of the acidic TME to circadian disorder, we subset TAM NSCLC patient samples into groups (Crem high TAMs and Crem low TAMs) based on median Crem expression. Notably, in macrophages from human NSCLC there was a trend toward disorder in Crem low but not Crem high TAM samples (Figure 7A,B). Additionally, the co-variance among core clock genes observed in alveolar macrophages from healthy donors was absent within Crem low and Crem high TAM samples (Figure 7C). In all, these data indicate that there is population-level disorder in the circadian rhythms of tumor-associated macrophages in humans and mice, suggesting that circadian rhythms are indeed altered in macrophages within the TME."

      And in the Discussion:

      "Indeed, we observed differences in the circadian clock of Crem low human TAM samples compared to Crem high human TAM samples, suggesting that acidic pH influences circadian disorder in TAMs (Figure 7). Interestingly, Crem low TAM samples exhibited a trend toward disorder while Crem high TAM samples did not. This is of particular interest, as we have observed that acidic pH can enhance circadian rhythms in macrophages, raising the question of whether acidic pH promotes or protects against circadian disorder."

      Minor comments: 1. Figure 2C needs clarification. It's unclear why pro-inflammatory macrophages treated with lactic acid would have a shorter amplitude and period, while acidic pH would increase amplitude and period in M2 macrophages.

      We thank Reviewer #1 for this important observation. Based on the comment, it is our understanding that the Reviewer is referring to the data in Figure 2 (low pH) compared to Figure 4 (lactate). We also find it very interesting that lactate alters rhythms in a manner distinct from the way in which acidic pH alters rhythms. Reviewer 3 asked for clarification on how lactate affected circadian gene expression in pH 7.4 or 6.5. We have added these data as Figure 4C (data and text below). It is notable that lactate opposing effects on circadian gene expression in pH 6.5, enhancing the effects of low pH in some cases (Nr1d1) while blunting them in other cases (Cry1). This is mentioned in the text.

      "Lactate was also observed to alter expression of the circadian clock genes Per2, Cry1, and Nr1d1 over time in BMDMs cultured at pH 6.5, while having more subtle effects at pH 7.4 (Figure 4C). Notably, lactate blunted the effect of pH 6.5 on Cry1 expression, while enhancing the effect of low pH on Nr1d1 expression."

      Why these two stimuli alter rhythms differently remains an open question that is discussed in the Discussion section and is prime to be a topic of future investigation. We have added to the Discussion section potential reasons why these conditions may alter rhythms differently, such as the different pathways downstream of sensing these two different conditions. Please see the updated text, below.

      "Although lactate polarizes macrophages toward a pro-resolution phenotype similar to acidic pH[30, 93], exposure to lactate had different effects on circadian rhythms - and in some cases, circadian clock gene expression - than exposure to acidic pH (Figure 4). Sensing of lactate occurs through different pathways than acid-sensing, which may contribute to the different ways in which these two stimuli modulate circadian rhythms of macrophages[111]. One previously published finding that may offer mechanistic insight into how phenotype can influence circadian rhythms is the suppression of Bmal1 by LPS-inducible miR-155[54]. It has also been observed that RORα-mediated activation of Bmal1 transcription is enhanced by PPARγ co-activation[112]. In macrophages, PPARγ expression is induced upon stimulation with IL-4 and plays a key role in alternative activation of macrophages, promoting a pro-resolution macrophage phenotype, and supporting resolution of inflammation[113-115]. Such observations prompt the question of whether there are yet-unidentified factors induced downstream of various polarizing stimuli that can modulate expression of circadian genes at the transcriptional and protein levels. Further work is required to understand the interplay between macrophage phenotype and circadian rhythms."

      The scale in Figure 2C should be equal for all conditions (e.g., -200).

      We appreciate Reviewer #1's preference for the axes to be scaled similarly to enable cross-comparison between graphs. However, due to the different amplitude of pro-inflammatory macrophages compared to the others, we feel that making all axes the same will make it hard to see the rhythms of pro-inflammatory macrophages, hindering the reader's ability to observe the data. Thus, we have put the matched-axis plots, shown below, in Supplementary Figure 4A.

      Absolute values of amplitude, damping, and period differ between Figure 1 and Figure 2A, B, C. The authors should explain these discrepancies.

      As with many experimental approaches, there is slight variation in absolute values between independent experiments, which Reviewer #1 correctly notes. However, while the absolute values vary slightly, the relationship between the values in each of these conditions remains the same across the panels mentioned by Reviewer #1.

      The authors should consider modulating the acidic environment of macrophages in settings more representative of cancer. For example, by adding conditioned media from tumor cells with pronounced glycolysis.

      We appreciate Reviewer #1's desire to more closely mimic the tumor microenvironment. To address Reviewer #1's point, we cultured macrophages in RPMI or cancer cell (KCKO) supernatant at pH 6.5 or pH-adjusted to pH 7.4 and assessed rhythms by measuring rhythmic activity of Per2-Luc with LumiCycle analysis. We then compared changes in rhythms between macrophages cultured normal media to cancer cell supernatant in pH-matched conditions to assess how cancer cell-conditioned media may influence circadian rhythms of macrophages, and the contribution of acidic pH. We have added these data, shown below, as a new Supplementary Figure 5, and included a discussion of these data in the manuscript. Please see the new Figure and updated text below.

      "Cancer cell supernatant alters circadian rhythms in macrophages in a manner partially reversed by neutralization of pH.

      We have observed that polarizing stimuli, acidic pH, and lactate can alter circadian rhythms. However, the tumor microenvironment is complex. Cancer cells secrete a variety of factors and deplete nutrients in the environment. To model this, we cultured BMDMs in RPMI or supernatant collected from KCKO cells, which are a murine model of pancreatic ductal adenocarcinoma (PDAC)[94, 95], at pH 6.5 or neutralized to pH 7.4 (Supplementary Figure 5). Circadian rhythms of BMDMs cultured in cancer cell supernatant at pH 7.4 or pH 6.5 exhibited increased amplitude and lengthened period compared to RPMI control at pH 7.4 or 6.5, respectively, indicating that cancer cell supernatant contains factors that can alter circadian rhythms of BMDMs. Notably, BMDMs cultured in cancer cell supernatant at pH 6.5 had increased amplitude and shortened period compared to BMDMs cultured in cancer cell-conditioned media at pH7.4, indicating that pH-driven changes in rhythms were maintained in BMDMs cultured in cancer cell supernatant. When the pH of cancer cell supernatant was neutralized to pH7.4, the increased amplitude was decreased, and the shortened period was lengthened, indicating that neutralizing acidic pH partially reverses the changes in rhythms observed in macrophages cultured in cancer cell supernatant at pH 6.5. These data further support our observations that acidic pH can alter circadian rhythms of macrophages both alone and in combination with various factors in the TME."

      And, in the Discussion:

      "We have shown that various stimuli can alter rhythms of macrophages in a complex and contributing manner, including polarizing stimuli, acidic pH, and lactate. TGFβ is produced by a variety of cells within the TME, and was recently identified as a signal that can modulate circadian rhythms[123, 124]. Additionally, when we exposed macrophages to cancer cell-conditioned media, rhythms were modulated in a manner distinct from acidic pH or lactate, with these changes in rhythms partially reversed by neutralization of the cancer cell-conditioned media pH (Supplementary Figure 5). It is conceivable that, in addition to acidic pH, other stimuli in the TME are influencing circadian rhythms to drive population-level disorder that we observed by CCD."

      Arg1 alone is not sufficient as an M2 polarization marker. The authors should include additional markers.

      We thank Reviewer #1 for bringing up this critical point in experimental rigor. While Arg1 is a commonly-used marker for M2 polarization, Reviewer #1 points out that polarization of macrophages is typically assessed by a full panel of markers characteristic of the M2 state. To address this point, we have expanded our panel to include several other markers of M2 polarization in mice such as Retnla, Ym1, MGL1, and CD206. In response to Reviewer 2's major point 2 and Reviewer 3's major point 4 below, we have also expanded our panel of markers used to assess the M1 polarization state with Tnfa, Il1b. and Il6. We have added these data, shown below, to Supplementary Figure 1 and updated the text appropriately. Please see the new Figure and updated text below.

      "Consistent with previous studies, we found that genes associated with anti-inflammatory and pro-resolution programming characteristic of IL-4 and IL-13-stimulated macrophages such as Arg1, Retnla, Chil3 (Ym1), Clec10a (MGL1), and Mrc1 (CD206) were induced in IL-4 and IL-13-stimulated macrophages, but not IFNγ and LPS-stimulated macrophages. In contrast, genes associated with pro-inflammatory activity characteristic of IFNγ and LPS-stimulated macrophages such as Nos2 (iNOS), Tnfa, Il1b, and Il6 were induced in IFNγ and LPS-stimulated macrophages, but not IL-4 and IL-13-stimulated macrophages (Supplementary Figure 1)[28, 30, 65, 71, 74, 75]. This indicates that macrophages stimulated with IL-4 and IL-13 were polarized toward a pro-resolution phenotype, while macrophages stimulated with IFNγ and LPS were polarized toward a pro-inflammatory phenotype."

      __ Significance__

      While the manuscript provides valuable insights and has obvious novelty, it requires a significant revision

      We thank Reviewer #1 for their deep read of our manuscript, and their helpful feedback and suggestions. As shown by the comments above, we are confident we have fully addressed each of the points that were made to result in a much-improved revised manuscript.

      __ Reviewer #2 __

      Evidence, reproducibility and clarity

      Knudsen-Clark et al. showed that the circadian rhythm of bone marrow-derived macrophages (BMDM) can be affected by polarization stimuli, pH of the microenvironment, and by the presence of sodium-lactate. Mechanistically, the acidic pH of cell microenvironment is partly regulated by intracellular cAMP-mediated signaling events in BMDM. The authors also showed that the circadian clock of peritoneal macrophages is also modified by the pH of the cell microenvironment. Using publicly available data, the authors showed that the circadian rhythm of tumor-associated macrophages is similar to that of Bmal1-KO peritoneal macrophages. In a murine model of pancreatic cancer, the authors showed that the tumor growth is accelerated in C57BL/6 mice co-injected with cancer cells and Bmal1-KO BMDM as compared to mice co-injected with cancer cell and wild type BMDM.

      We thank Reviewer #2 for their insightful and helpful comments and feedback. Their Review guided key clarifying experiments and additions to the Discussion and Methods. To summarize, we added new data to Supplementary Figure 1 to characterize distinct gene expression in our different polarized macrophage populations, showed in Supplementary Figure 2 that serum shock independently induces cAMP and Icer, discussed the limitations of the artificial polarization models more clearly, and updated our Methods to better explain how macrophages were isolated from the peritoneum. We also quantified multiple immunoblots of pCREB, provided clarity in the Methods and Reviewer-only data on how our protein-extraction protocol isolates nuclear protein, better introduced the BMAL1-KO mouse model, and showed in Supplementary Figure 6 that low pH can induce oscillations in the absence of a serum shock.

      Major points of criticism: 1. Nine main figures include different experimental models on a non-systematic manner in the manuscript, and only literature-based correlation is used to link the results each other. The authors used in vitro BMDM and peritoneal cell-based model systems to study the effects of IL4+IL13, IFNg+LPS, low pH, sodium-lactate, adenylate cyclase inhibitors on the circadian clock of macrophages. The link between these microenvironment conditions of the cells is still correlative with the tumor microenvironment: publicly available data were used to correlate the increased expression level of cAMP-activated signaling events with the presence of acidic pH of tumor microenvironment. Notably, the cell signaling messenger molecule cAMP is produced by not only low extracellular pH by activated GPCRs, but also starvation of the cell. The starvation is also relevant to this study, since the BMDM used in the in vitro culture system were starving for 24 hours before the measurement of Per2-Luc expression to monitor circadian rhythm.

              We agree with the important point that Reviewer #2 makes that our synchronization protocol of serum starvation followed by serum shock can impact the cAMP signaling pathway. Indeed, it has previously been shown that serum shock induces phosphorylation of CREM in rat fibroblasts, which is indicative of signaling through the cAMP pathway. To address this point, we have added a schematic of our synchronization protocol to Supplementary Figure 2B for additional clarity. We have also performed additional experiments to test whether cAMP signaling is induced in macrophages by our synchronization protocol. For this, we assessed downstream targets of the cAMP signaling pathway, Icer and pCREB, after serum starvation but before serum shock, and at several time points post-treatment with serum shock (Supplementary Figures 2D,E). We observed that Icer and phosphorylation of Creb are induced rapidly in macrophages upon exposure to serum shock, as early as 10 minutes for pCREB and 1 hour post-exposure for Icer. Notably, this signaling is transient and rapidly returns to baseline, with pCREB levels fully returned to baseline by 2 hours post-treatment, at which time media is replaced and the experiment begins (CT 0). These data, shown below, have been added to Supplementary Figure 2 and a discussion of these data has been added to the manuscript - please see the modified text below.
      

      "The synchronization protocol we use to study circadian rhythms in BMDMs involves a 24-hour period of serum starvation followed by 2 hours of serum shock. It has previously been shown that serum shock can induce signaling through the cAMP pathway in rat fibroblasts[98]. To determine whether the synchronization protocol impacts cAMP signaling in macrophages, we harvested macrophages before and after serum shock. We then assessed Icer expression and phosphorylation of cyclic AMP-response element binding protein (CREB), which occur downstream of cAMP and have been used as readouts to assess induction of cAMP signaling in macrophages[29, 96, 100]. Serum shock of macrophages following serum starvation led to rapid phosphorylation of CREB and Icer expression that quickly returned to baseline (Supplementary Figure 2D,E). This indicates that serum starvation followed by serum shock in the synchronization protocol we use to study circadian rhythms in BMDMs induces transient signaling through the cAMP signaling pathway. "

      The definition of pre-resolution macrophages (MF) used across the manuscript could be argued. The authors defined BMDM polarized with IL-4 and IL-13 as pre-resolution MF. Resolution is followed by inflammation, but the IL-4 secretion does not occur in every inflammatory setting. Moreover, IL-4 and IL-13 are secreted during specific tissue environment and immunological settings involving type 2 inflammation or during germinal center reactions of the lymph nodes. • What are the characteristics of pre-resolution macrophages (MF)? The authors indicated that IL-4 and IL-13 cytokines were used to model the pre-resolution macrophages. In which pathological context are these cytokines produced and induce pre-resolution macrophages? IL-4 polarized BMDM can also produce pro-inflammatory protein and lipid mediators as compared to LPS-stimulated BMDM, and IL-4 polarized BMDM still have potent capacity to recruit immune cells and to establish type 2 inflammation.

      • The authors showed Arg1 and Vegfa qPCR data from BMDM only. Based on the literature, these MFs are anti-inflammatory cells particularly. Resolution-related MFs followed by acute inflammation are a specific subset of MFs, and the phenotype of pre-resolution MF should be described, referred, and measured specifically.

      We thank Reviewer #2 for bringing up this important point that clarity is required in describing our in vitro macrophage models. We chose the most commonly used models of in vitro macrophage polarization in the tumor immunology field, M2 (IL-4+IL-13) and M1 (IFNγ+LPS). These polarization conditions have been used for over two decades in the field, and have been well-characterized to drive a pro-inflammatory (for M1) and pro-resolution or anti-inflammatory (for M2) macrophage phenotype (Murray 2017 Annu Rev Phys). Each of these cell states have similarities in phenotype to pro-inflammatory and pro-resolution (pro-tumorigenic) macrophages found in tumors. In fact, in the literature, pro-inflammatory and pro-resolution TAMs will frequently be categorized as "M1" or "M2", respectively, even though this is a gross oversimplification (Ding 2019 J Immunol, Garrido-Martin 2020 J Immunother Cancer).

      As Reviewer #2 points out, IL-4 and IL-13 play a role in inflammatory settings, mediating protective responses to parasites and pathological responses to allergens. Importantly, IL-4 and IL-13 are also key regulators and effectors of resolution and wound repair (Allen 2023 Annu Rev Immunol). In line with this, M2 macrophages show many of the characteristics of pro-resolution programming in their gene expression profile, expressing genes associated with wound healing (ex. Vegf) and immunoregulation (ex. Arg1) (Mantovani 2013 J Pathol). These cells have frequently been used as a model for studying TAMs in vitro, due to the similarity in pro-resolution programming that is dysregulated/hijacked in TAMs (Biswas 2006 Blood). M2 macrophages have also been referred to as anti-inflammatory, and this is in line with their role in the type 2 response driven by IL-4 and IL-13, as this is primarily a response induced by allergy or parasites where tissue damage drives an anti-inflammatory and pro-resolution phenotype in macrophages (Pesce 2009 Plos Pathogens and Allen 2023 Annu Rev Immunol).

      We do not assert that these in vitro models recapitulate the macrophage polarization cycle that Reviewer #2 astutely describes, and indeed, stimuli polarizing macrophages in tumor are much more diverse and complex (Laviron 2022 Cell Rep). We also fully agree with Reviewer #2 that, while IL4 and IL13 may exist in the tumor and be secreted by Th2 CD4 T cells (see Shiao 2015 Cancer Immunol Res), there may be multiple reasons why macrophages may be polarized to a pro-resolution, M2-like state in a tumor (in fact, exposure to low pH and lactate each independently do this, as we show in Supplementary Figure 2 and Figure 4, and was previously shown in Jiang 2021 J Immunol and Colegio 2014 Nature). Nonetheless, using the well-described M1 and M2 in vitro models allows our findings to be directly comparable to the vast literature that also uses these models, and to understand how distinct polarization states respond to low pH.

      We fully agree with Reviewer #2 that these cells must be defined more clearly in the text. We have taken care to discuss the limitations of using in vitro polarization models to study macrophages in our Limitations of the Study section. To better address Reviewer #2's concern, we have more thoroughly introduced the M2 macrophages as a model, and are clear that that these are type 2-driven macrophages that share characteristics of pro-resolution macrophages. We have also added additional citations to the manuscript, including those highlighted above in our response. Finally, we have expanded our panel to better characterize the IL-4/IL-13 stimulated macrophages using more markers that have been characterized in the literature, in line with both Reviewer #2's comments and that of Reviewer #1 and Reviewer #3. Please see the updated data and text, below.

      "As macrophages are a phenotypically heterogeneous population in the TME, we first sought to understand whether diversity in macrophage phenotype could translate to diversity in circadian rhythms of macrophages. To this end, we used two well-established in vitro polarization models to study distinct macrophage phenotypes[5, 60-63]. For a model of pro-inflammatory macrophages, we stimulated macrophages with IFNγ (interferon γ) and LPS (lipopolysaccharide) to elicit a pro-inflammatory phenotype[60, 64]. These macrophages are often referred to as 'M1' and are broadly viewed as anti-tumorigenic, and we will refer to them throughout this paper as pro-inflammatory macrophages[65, 66]. For a model at the opposite end of the phenotypic spectrum, we stimulated macrophages with IL-4 and IL-13[60, 67]. While these type 2 stimuli play a role in the response to parasites and allergy, they are also major drivers of wound healing; in line with this, IL-4 and IL-13-stimulated macrophages have been well-characterized to adopt gene expression profiles associated with wound-healing and anti-inflammatory macrophage phenotypes[68-71]. As such, these macrophages are often used as a model to study pro-tumorigenic macrophages in vitro and are often referred to as 'M2' macrophages; throughout this paper, we will refer to IL-4 and IL-13-stimulated macrophages as pro-resolution macrophages[66, 72, 73]. Consistent with previous studies, we found that genes associated with anti-inflammatory and pro-resolution programming characteristic of IL-4 and IL-13-stimulated macrophages such as Arg1, Retnla, Chil3 (Ym1), Clec10a (MGL1), and Mrc1 (CD206) were induced in IL-4 and IL-13-stimulated macrophages, but not IFNγ and LPS-stimulated macrophages. In contrast, genes associated with pro-inflammatory activity characteristic of IFNγ and LPS-stimulated macrophages such as Nos2 (iNOS), Tnfa, Il1b, and Il6 were induced in IFNγ and LPS-stimulated macrophages, but not IL-4 and IL-13-stimulated macrophages (Supplementary Figure 1)[28, 30, 65, 71, 74, 75]. This indicates that macrophages stimulated with IL-4 and IL-13 were polarized toward a pro-resolution phenotype, while macrophages stimulated with IFNγ and LPS were polarized toward a pro-inflammatory phenotype.

      In the Limitations of the Study section, we now write the following:

      "Our observations of rhythms in macrophages of different phenotypes are limited by in vitro polarization models. It is important to note that while our data suggest that pro-inflammatory macrophages have suppressed rhythms and increased rate of desynchrony, it remains unclear the extent to which these findings apply to the range of pro-inflammatory macrophages found in vivo. We use IFNγ and LPS co-treatment in vitro to model a pro-inflammatory macrophage phenotype that is commonly referred to as 'M1', but under inflammatory conditions in vivo, macrophages are exposed to a variety of stimuli that result in a spectrum of phenotypes, each highly context-dependent. The same is true for for 'M2'; different tissue microenvironment are different and pro-resolution macrophages exist in a spectrum."

      The authors used IFNg and LPS, or IL-4 and IL-13 and co-treatments to polarize BMDM in to type 1 (referred as pro-inflammatory MF) and type 2 (referred as pre-resolution MF) activation state. The comparison between these BMDM populations has limitations, since LPS induces a potent inflammatory response in MF. The single treatment with MF-polarizing cytokines enable a more relevant comparison to study the circadian clock in classically and alternatively activated MF.

      We thank Reviewer #2 for bringing up this important point to provide additional clarity on our polarization conditions. The use of IFNγ and LPS to polarize macrophages toward a pro-inflammatory, M1 phenotype, and the use of IL-4 an IL-13 to polarize macrophages toward a pro-resolution, M2 phenotype have been commonly used for over two decades, and thus are well-characterized in the literature (please see Murray 2017 Annu Rev Phys for an extensive review on the history of these polarization models, as well as Hörhold 2020 PLOS Computational Biology, Binger 2015 JCI, McWhorter 2013 PNAS, Ying 2013 J Vis Exp for more recent studies using these models). The use of LPS alone or in combination with IFNγ, and IL-13 along with IL-4, was introduced in 1998 (Munder 1998 J Immunol). This approach was originally designed to mimic what could happen when macrophages were exposed to CD4+ Th1 cells, which produce IFNγ, or Th2 cells, which produce IL-4 and IL-13 (Munder 1998 J Immunol, Murray 2017 Annu Rev Phys). As Reviewer #2 points out, these stimuli induce potent responses, driving macrophages to adopt pro-inflammatory or pro-resolution/anti-inflammatory phenotypes that are two extremes at opposite ends of the spectrum of macrophage phenotypes (Mosser 2008 Nat Rev Immunol). Since our goal was to study rhythms of distinct macrophage phenotypes in vitro, and how TME-associated conditions such as acidic pH and lactate affect their rhythms, these cell states were appropriate for our questions. Thus, the polarization models used in this paper allowed us to achieve this goal. We include a section in the Discussion on the limitations of in vitro polarization models.

      "A critical question in understanding the role of circadian rhythms in macrophage biology is determining how different polarization states of macrophages affect their internal circadian rhythms. This is especially important considering that tumor-associated macrophages are a highly heterogeneous population. Our data indicate that compared to unstimulated macrophages, rhythms are enhanced in pro-resolution macrophages, characterized by increased amplitude and improved ability to maintain synchrony; in contrast, rhythms are suppressed in pro-inflammatory macrophages, characterized by decreased amplitude and impaired ability to maintain synchrony (Figure 1). These agree with previously published work showing that polarizing stimuli alone and in combination with each other can alter rhythms differently in macrophages[80, 81]. In a tumor, macrophages exist along a continuum of polarization states and phenotypes[18-21, 24]. Thus, while our characterizations of rhythms in in vitro-polarized macrophages provide a foundation for understanding how phenotype affects circadian rhythms of macrophages, further experiments will be needed to assess macrophages across the full spectrum of phenotypes. Indeed, alteration of rhythms may be just as highly variable and context-dependent as phenotype itself."

      There are missing links between the results of showing the circadian rhythm of polarized BMDM, sodium-lactate treated BMDM, and tumor growth. Specifically, do the used pancreatic ductal adenocarcinoma cells produce IL-4 and sodium-lactate? In the LLC-based experimental in silico analysis of tumors, the LLC do not produce IL-4.

      Reviewer #2 raises important points about the source of lactate and IL-4 in tumors as relevance for our investigation of how these factors can alter rhythms in macrophages. Tumor-infiltrating Th2 CD4 T cells are potential sources of IL-4 and IL-13 in the tumor (see Shiao 2015 Cancer Immunol Res). Various cells in the tumor can produce lactate. We discuss this in both the Introduction and the Results: poor vascularization of tumors results in hypoxia areas, where cells are pushed toward glycolysis to survive and thus secrete increased glycolytic waste products such as protons and lactate. As lactate is lactic acid, ionized it is sodium l-lactate.

      How can the circadian rhythm affect the function of BMDM? The Authors should provide evidence that circadian rhythm affects the function of polarized MF.

      We agree with Reviewer #2 that the next step is to determine how altered rhythms influence function of macrophages. This will be the topic of future work, but is outside the scope of this paper. Our contribution with this paper is providing the first evidence that rhythms are altered in the TME and the TME-associated conditions can alter rhythms in macrophages. We have added what is currently known about how circadian rhythms influence macrophages function to the discussion section to facilitate a conversation about this important future direction. Please see the updated text below.

      "Considering our observations that conditions associated with the TME can alter circadian rhythms in macrophages, it becomes increasingly important to understand the relevance of macrophage rhythms to their function in tumors. It has been shown that acidic pH and lactate can each drive functional polarization of macrophages toward a phenotype that promotes tumor growth, with acidic pH modulating phagocytosis and suppressing inflammatory cytokine secretion and cytotoxicity[28, 30, 93]. However, how the changes in circadian rhythms of macrophages driven by these conditions contributes to their altered function remains unknown. Current evidence suggests that circadian rhythms confer a time-of-day-dependency on macrophage function by gating the macrophage response to inflammatory stimuli based on time-of-day. As such, responses to inflammatory stimuli such as LPS or bacteria are heightened during the active phase while the inflammatory response is suppressed during the inactive phase. An important future direction will be to determine how changes in circadian rhythms of macrophages, such as those observed under acidic pH or high lactate, influences the circadian gating of their function."

      In Figure 3, the authors show data from peritoneal cells. The isolated peritoneal cells are not pure macrophage populations. Based on the referred article in the manuscript, the peritoneal cavity contains more then 50% of lymphocytes, and the myeloid compartment contains 80% macrophages.

      Reviewer #2 raises important concerns about the purity of the peritoneal population used in our experiments. We enrich for peritoneal macrophages from the peritoneal exudate cells by removing non-adherent cells in culture. This is described in our Methods section and is a method of isolation that is commonly used in the field, as lymphocytes are non-adherent. In addition to the source cited in the paper within our Methods section (Goncalves 2015 Curr Prot Immunol), please see Layoun 2015 J Vis Exp, de Jesus 2022 STAR Protocols, and Harvard HLA Lab protocol - macrophages enriched in this manner have been shown to be over 90% pure. We have modified our Methods section to make this clear, and added the additional references in this response to this section of our Methods. Please see the modified text below.

      "Peritoneal exudate cells were harvested from mice as previously published[137]. To isolate peritoneal macrophages, peritoneal exudate cells were seeded at 1.2*106 cells/mL in RPMI/10% HI FBS supplemented with 100U/mL Penicillin-Streptomycin and left at 37⁰C for 1 hour, after which non-adherent cells were rinsed off[136]. Isolation of peritoneal macrophages using this method has been shown to yield a population that is over 90% in purity[138, 139]. Peritoneal macrophages were then cultured in Atmospheric Media at pH 7.4 or 6.5 with 100μM D-luciferin, and kept at 37⁰C in atmospheric conditions."

      The figure legend of Figure 3 describes the effects of pH on the circadian rhythm of bone marrow-derived macrophages ex vivo. Peritoneal macrophages involve tissue resident peritoneal macrophages with yolk sac and fetal liver origin, and also involve small peritoneal MF with bone marrow origin. The altered description of results and figure legends makes confusion.

      We are very grateful to Reviewer #2 for pointing out our typo. We have fixed the caption of Figure 3 to properly describe the data as "peritoneal macrophages ex vivo".

      In Figure 6C, one single Western blot is shown with any quantification. The authors should provide data of the relative protein level of p-CREB from at least 3 independent experiments. In the Western-blot part of the methods, the authors described that the pellet was discarded after cell lysis. The p-CREB is the activated form of the transcription factor CREB and there is increased binding to the chromatin to regulate gene expression. By discarding the pellet after cell lysis, the chromatin-bond p-CREB could be also removed at the same time.

      We thank Reviewer 2 for bringing up this point. We agree that quantification is an important aspect of western blot. We have repeated the experiment again for n=3 and provide quantification of pCREB normalized to total protein. We have added these data, shown below, to Figure 5.

      Reviewer #2 also expressed concern that we may not be capturing all of the CREB due to nuclear localization and chromatin binding. We specifically chose the lysis buffer M-Per, which is formulated to lyse the nucleus and solubilize nuclear and chromatin-bound proteins. To demonstrate this, we show in the below Figure to the Reviewer that the nuclear protein p85 is solubilized and readily detectable by western blot using our protein extraction method.

      We have also added an additional sentence in the Methods section for clarity - please see the modified text below.

      "Cells were lysed using the M-Per lysis reagent (Thermo Scientific, CAT#78501), supplemented with protease and phosphatase inhibitor cocktail (1:100; Sigma, CAT#PPC1010) and phosphatase inhibitor cocktail 2 (1:50; Sigma, CAT#P5726), with 200μM deferoxamine (Sigma, CAT#D9533). M-Per is formulated to lyse the nucleus and solubilize nuclear and chromatin-bound proteins, allowing isolation of nuclear proteins as well as cytosolic proteins. Lysates were incubated on ice for 1 hour, then centrifuged at 17,000 xg to pellet out debris; supernatant was collected."

      It is confusing that adenylate-cyclase inhibitor MDL-12 elevated the phospho-CREB levels in BMDM. How can the authors exclude any other inducers of CREB phosphorylation?

      We agree with Reviewer #2 that it is surprising pCREB was elevated with MDL-12 treatment alone, and we do indeed think that there are other pathways contributing to this. We have addressed this point in the Discussion - please see the text below.

      "The mechanism through which acidic pH can modulate the circadian clock in macrophages remains unclear. Evidence in the literature suggests that acidic pH promotes a pro-resolution phenotype in macrophages by driving signaling through the cAMP pathway[29]. It has previously been shown that cAMP signaling can modulate the circadian clock[99]. However, our data indicated that cAMP signaling was not fully sufficient to confer pH-mediated changes in circadian rhythms of macrophages (Figure 5A,B). Treatment with MDL-12, commonly known as an inhibitor of adenylyl cyclase[29, 117], resulted in suppression of pH-induced changes in amplitude of circadian rhythms but did not inhibit signaling through the cAMP signaling pathway (Figure 5C,D). While MDL-12 is commonly used as an adenylyl cyclase inhibitor, it has also been documented to have inhibitory activity toward phosphodiesterases (PDEs) and the import of calcium into the cytosol through various mechanisms[118, 119]. This is of particular interest, as calcium signaling has also been shown to be capable of modulating the circadian clock[120]. Furthermore, while acid-sensing through GPCRs have been the most well-characterized pathways in macrophages, there remain additional ways in which acidic pH can be sensed by macrophages such as acid-sensing ion channels[121, 122]. Further work is required to understand the signaling pathways through which pH can influence macrophage phenotype and circadian rhythms."

      It is described in the methods that BMDM were starving for 24 hours in serum-free culture media followed by serum shock (50% FBS). The cAMP production can be induced during cell starvation which should be considered for the data representation.

      We appreciate that Reviewer #2 points out that our synchronization protocol of serum starvation followed by serum shock may impact the cAMP signaling pathway in macrophages, as serum shock has been shown to induce pCREB, a downstream mediator of cAMP signaling, in rat fibroblasts. Indeed, we show in additional experiments performed (in response to Reviewer #2's major comment 1) evidence that cAMP signaling is induced in macrophages following the serum shock phase of our synchronization protocol, as indicated by elevation of Icer and pCREB. As we note above, this induction is transient and returns to baseline by 2 hours post-serum shock, the time at which we replace media and begin our experiments (CT 0).

      Despite the transient nature of cAMP induction by our synchronization protocol, we agree wholeheartedly with Reviewer #2 that this must be considered in light of our experimental system in which we are studying the effect of acidic pH on circadian rhythms of macrophages, which in itself induces signaling through the cAMP signaling pathway. To address Reviewer #2's point, we have performed experiments in which we culture unstimulated BMDMs in neutral pH 7.4 or acidic pH 6.5, without prior serum starvation and serum shock (i.e. we do not submit these BMDMs to the synchronization protocol). We then observed circadian rhythms of Per2-Luc by LumiCycle to determine whether acidic pH alters circadian rhythms of BMDMs in the absence of prior serum starvation followed by serum shock. Similar to our observations in Figure 2, circadian rhythms of macrophages at pH 6.5 had increased amplitude and shortened period compared to rhythms of macrophages at pH 7.4. This indicates that pH-driven changes in circadian rhythms observed in our system are not due to the synchronization protocol. The data, shown below, have been placed in a new Supplementary Figure 6, and a discussion of these results has been added to the Results section - please see the updated text below.

      "As acidic pH induces signaling through the cAMP pathway, we sought to determine whether acidic pH independently contributed to the pH-driven changes in circadian rhythms we observe in BMDMs. To test this, we omitted the synchronization step and observed BMDM rhythms by LumiCycle when cultured in neutral pH 7.4 or acidic pH 6.8 or pH 6.5 (Supplementary Figure 6). Circadian rhythms of BMDMs cultured at pH 6.5 exhibited similar changes as previously observed, with enhanced amplitude and shortened period relative to BMDMs at pH 7.4. This indicates pH-driven changes observed in circadian rhythms of BMDMs occur in the absence of prior serum starvation and serum shock. "As acidic pH independently induces signaling through the cAMP pathway, we sought to determine whether acid pH could also independently contribute to the pH-driven changes in circadian rhythms we observe in BMDMs. To test this, we omitted the synchronization step and observed BMDM rhythms by LumiCycle when cultured in neutral pH 7.4 or acidic pH 6.8 or pH 6.5 (Supplementary Figure 6). Circadian rhythms of BMDMs cultured at pH 6.5 exhibited similar changes as previously observed, with enhanced amplitude and shortened period relative to BMDMs at pH 7.4. This indicates pH-driven changes observed in circadian rhythms of BMDMs occur in the absence of prior serum starvation and serum shock."

      How can the authors explain and prove that the wild type and Bmal1-KO BMDM co-injected with pancreatic cancer cells subcutaneously survive, present, and have effector functions at the same extent in the subcutaneous tissue, before and during tumor growth (Figure 9)? In other words, what kind of MF-derived parameters could be modified by disrupting the circadian rhythm of MF during tumor development? The production of MF-derived regulatory enzymes, cytokines, growth factors are affected by the disrupted circadian clock in MF?

              Review #2 poses the very important question of why we see differences in tumor growth in our co-injection model, and what might be driving it. Of note, co-injection models of tumor growth are commonly used to determine macrophage-specific roles in tumor growth (Colegio 2014 Nature, Mills 2019 Cell Rep, Lee 2018 Nat Comm). We observed that tumor growth is altered when macrophages with disrupted circadian rhythms (BMAL1 KO) are co-injected compared to when macrophages with intact circadian rhythms (WT) are co-injected in a murine model of pancreatic cancer using KCKO cells. Our observation is supported by a previously published paper in which they used a co-injection model of melanoma, which we cite in the manuscript(Alexander 2020 eLife). What drives this difference in tumor growth remains an open question that is the subject of future work and is outside the scope of this paper, which focuses on our discovery that factors associated with the tumor microenvironment can alter circadian rhythms in macrophages. We have included a discussion on what is currently known about how circadian rhythms alter macrophage function, acknowledging that we have yet to answer these important questions and identifying it as interest for future work. Please see the text below.
      

      "Considering our observations that conditions associated with the TME can alter circadian rhythms in macrophages, it becomes increasingly important to understand the relevance of macrophage rhythms to their function in tumors. It has been shown that acidic pH and lactate can each drive functional polarization of macrophages toward a phenotype that promotes tumor growth, with acidic pH modulating phagocytosis and suppressing inflammatory cytokine secretion and cytotoxicity[28, 30, 93]. However, how the changes in circadian rhythms of macrophages driven by these conditions contributes to their altered function remains unknown. Current evidence suggests that circadian rhythms confer a time-of-day-dependency on macrophage function by gating the macrophage response to inflammatory stimuli based on time-of-day. As such, responses to inflammatory stimuli such as LPS or bacteria are heightened during the active phase while the inflammatory response is suppressed during the inactive phase. An important future direction will be to determine how changes in circadian rhythms of macrophages, such as those observed under acidic pH or high lactate, influences the circadian gating of their function. Data from our lab and others suggest that disruption of the macrophage-intrinsic circadian clock accelerates tumor growth, indicating that circadian regulation of macrophages is tumor-suppressive in models of PDAC (our work) and melanoma [109]. This agrees with complementary findings that behavioral disruption of circadian rhythms in mice (through chronic jetlag) disrupts tumor macrophage circadian rhythms and accelerates tumor growth[56]. It remains unclear whether this is through the pro-tumorigenic functions of macrophages such as extracellular matrix remodeling or angiogenesis, through suppression of the anti-tumor immune response, or a combination of both functions. Further work will be needed to tease apart these distinctions."

      Minor points of criticism: 1. The figure legends of the graphs and diagrams are missing in Figure 2D,E,F

      We thank Reviewer #2 for pointing out that figure legends were missing. We have added legends for Figure 2D,E,F.

      The BMAL1-based in vivo murine model of circadian rhythm is not introduced in the manuscript.

      We thank Reviewer #2 for bringing to our attention that the BMAL1 KO macrophage model was not well-introduced in the manuscript. To address this point, we have modified the text to better introduce this model. Please see the modified text below.

      "As a positive control for circadian clock disruption, we used data from BMAL1 KO peritoneal macrophages [44]. BMAL1 KO macrophages have a genetic disruption of the circadian clock due to the loss of Bmal1, the central clock gene. As a result, circadian rhythms of BMAL1 KO macrophages are disrupted, lacking rhythmicity and downstream circadian regulation of macrophage function (Supplementary Figure 8)[45, 54]. "As a positive control for circadian clock disruption, we used data from BMAL1 KO peritoneal macrophages [44]. BMAL1 KO macrophages have a genetic disruption of the circadian clock due to the loss of Bmal1, the central clock gene. As a result, circadian rhythms of BMAL1 KO macrophages are disrupted, lacking rhythmicity and downstream circadian regulation of macrophage function (Supplementary Figure 8)[45, 54]."__ __

      Significance

      Knudsen-Clark et al. showed that the circadian rhythm of bone marrow-derived macrophages (BMDM) can be affected by polarization stimuli, pH of the microenvironment, and by the presence of sodium-lactate. Mechanistically, the acidic pH of cell microenvironment is partly regulated by intracellular cAMP-mediated signaling events in BMDM. The authors also showed that the circadian clock of peritoneal macrophages is also modified by the pH of the cell microenvironment. Using publicly available data, the authors showed that the circadian rhythm of tumor-associated macrophages is similar to that of Bmal1-KO peritoneal macrophages. In a murine model of pancreatic cancer, the authors showed that the tumor growth is accelerated in C57BL/6 mice co-injected with cancer cells and Bmal1-KO BMDM as compared to mice co-injected with cancer cell and wild type BMDM.

      We are grateful to Reviewer #2 for their very helpful comments and suggestions, which we believe have greatly enhanced the clarity and reproducibility of this manuscript.

      __Reviewer #3 (Evidence, reproducibility and clarity (Required)): __

      Review for Knudsen-Clark et al.

      "Circadian rhythms of macrophages are altered by the acidic pH of the tumor microenvironment"

      Knudsen-Clark and colleagues explore the impact of TME alterations on macrophage circadian rhythms. The authors find that both acidic pH and lactate modulate circadian rhythms which alter macrophage phenotype. Importantly, they define circadian disruption of tumor-associated macrophages within the TME and show that circadian disruption in macrophages promotes tumor growth using a PDAC line. This represents an important understanding of the crosstalk between cancer cells and immune cells as well as the understanding of how the TME disrupts circadian rhythms. The study is well-done, however, authors need to address several important points below.

      We thank Reviewer #3 for their in-depth and insightful comments and suggestions, which have resulted in a much-improved manuscript. We were pleased that Reviewer #3 found the work to be "an important study that is well-done" and that it "represents an important understanding of the crosstalk between cancer cells and immune cells as well as the understanding of how the TME disrupts circadian rhythms.". In response to Reviewer #3's comments, we have added several new key experiments and changes to the text. To summarize, we added new data to Supplementary Figure 1 to better characterize our macrophage polarization states, showed in Figure 3 that low pH affects peritoneal macrophage circadian gene expression in a similar fashion as bone marrow-derived macrophages, added new data in Figure 4 to show how lactate and low pH affect circadian gene expression over time, and new computational analysis to Figures 6, 7, and Supplementary Figure 9 to probe circadian gene covariance from publicly available data. We also made several key additions to the Discussion to discuss the functional implications of macrophage circadian rhythm disruption by low pH and potential mechanisms of this disruption. Finally, at the request of Reviewer #3, we consolidated several existing Figures and added new data, where appropriate, to existing figures, and we worked to describe new findings succinctly.

      Major comments:

      • In Figures 3 and 4, the authors can include additional clock genes that can be run by qPCR. This was done in Figure 2 and was a nice addition to the data.

      We agree with Reviewer #3's suggestion that an analysis of clock gene expression at the mRNA level would enhance our data in Figures 3 and 4. To address this point, we have performed short time course experiments to assess circadian clock gene expression over time in BMDMs cultured with or without lactate at neutral or acidic pH (for Figure 4). In line with the difference in circadian rhythms of Per2-Luc levels between BMDMs cultured in the presence or absence of lactate which we observed by Lumicycle analysis, we measured changes in expression of the circadian clock genes Per2, Nr1d1, and Cry1 between macrophages cultured with 25 mM sodium-L-lactate compared to those cultured with 0 mM sodium-L-lactate at pH 6.5. We have added these data, shown below, to Figure 4, and updated the manuscript accordingly to discuss these results. Please see below for the new Figure Panel and modified text.

      "Lactate was also observed to alter expression of the circadian clock genes Per2, Cry1, and Nr1d1 over time in BMDMs cultured at pH 6.5, while having more subtle effects at pH 7.4 (Figure 4C). Notably, lactate blunted the effect of pH 6.5 on Cry1 expression, while enhancing the effect of low pH on Nr1d1 expression. In all, these data indicate that concentration of lactate similar to that present in the TME can influence circadian rhythms and circadian clock gene expression of macrophages."

      As an additional measure to address Reviewer #3's point about Figure 3 (peritoneal macrophages), we have compared expression of circadian clock genes in peritoneal macrophages cultured at neutral pH 7.4 or acidic pH 6.8 for 24 hours using a publicly available RNA-seq data set from Jiang 2021 J Immunol (GSE164697). In line with previous observations in macrophages cultured under acidic compared to neutral pH conditions, including the clock gene expression data from Figure 2 in BMDMs and the Per2-Luc levels observed in peritoneal macrophages in Figure 3, we found that peritoneal macrophages exhibited differences in expression of circadian clock genes when cultured at acidic pH 6.8 compared to neutral pH 7.4. We have added these data, shown below, as Figure 3B, and have updated the manuscript accordingly - please see below for the new Figure panel and modified text.

      "To test whether pH-driven changes in circadian rhythms of peritoneal macrophages were reflected at the mRNA level, we compared expression of circadian clock genes in peritoneal macrophages cultured at neutral pH 7.4 or acidic pH 6.8 for 24 hours using publicly available RNA-sequencing data [30]. In line with altered circadian rhythms observed by Lumicycle, peritoneal macrophages cultured at pH 6.8 expressed different levels of circadian clock genes than peritoneal macrophages culture at pH 7.4 (Figure 3B). The trends in changes of gene expression in peritoneal macrophages cultured at pH 6.8 matched what we observed in BMDMs, where low pH generally led to higher levels of circadian clock gene expression (Figure 2D-F). These data support our observations by LumiCycle and indicate that acidic pH drives transcriptional changes in multiple components of the circadian clock. In all, these data are evidence that pH-dependent changes in circadian rhythms are relevant to in vivo-differentiated macrophages."

      We have also updated the Methods section appropriately

      "FASTQ files from a previously published analysis of peritoneal macrophages cultured under neutral pH 7.4 or acidic pH 6.8 conditions were downloaded from NCBI GEO (accession #GSE164697) [30]."

      2) There are far too many figures with minimal data in each. Please consolidate the figures. For example, Figures 1-3 can be fully combined, Figures 4-6 can be combined, and Figures 7-8 can be combined. Additionally, it is unclear if Figure 5 needs to be in the main, it can be moved to the supplement.

      We appreciate the preference of Reviewer #3 to see some of the figures consolidated. We have combined Figures 5 and 6 into a single new Figure 5. Additionally, we have added new data from revisions to current figures to increase the amount of data in each figure and minimize the amount of new figures generated. In all, despite the large amount of new data added to the paper in response to Reviewer comments and suggestions (including additional data in Figure 4 and new Figures 6 and 8), our manuscript now contains 10 main Figures, only one more than the initial submission.

      3) The observation that conditions like pH and lactate alter macrophage phenotype and rhythmicity are important. However, macrophage phenotype via gene expression does not always correlate to function. It is important for authors to demonstrate the effect of pH or lactate on macrophage function. This can be done using co-culture assays with cancer cells. If these experiments cannot be performed, it is suggested that authors discuss these limitations and consideration in the discussion.

      Reviewer #3 correctly points out that changes in phenotype does not always correlate to changes in function. Others have shown that acidic pH and lactate can each alter macrophage phenotype, and also alter macrophage function and the ability to promote tumor growth (please see El-Kenawi 2019 Br J Cancer, Jiang 2021 J Immunol, Colegio 2014 Nature). How changes in rhythms influence macrophage function remains unknown and we agree with Reviewer #3 that this is an important future direction, We have added a section in the Discussion to facilitate the discussion of this important future direction. Please see the text below.

      "Considering our observations that conditions associated with the TME can alter circadian rhythms in macrophages, it becomes increasingly important to understand the relevance of macrophage rhythms to their function in tumors. It has been shown that acidic pH and lactate can each drive functional polarization of macrophages toward a phenotype that promotes tumor growth, with acidic pH modulating phagocytosis and suppressing inflammatory cytokine secretion and cytotoxicity[28, 30, 93]. However, how the changes in circadian rhythms of macrophages driven by these conditions contributes to their altered function remains unknown. Current evidence suggests that circadian rhythms confer a time-of-day-dependency on macrophage function by gating the macrophage response to inflammatory stimuli based on time-of-day. As such, responses to inflammatory stimuli such as LPS or bacteria are heightened during the active phase while the inflammatory response is suppressed during the inactive phase. An important future direction will be to determine how changes in circadian rhythms of macrophages, such as those observed under acidic pH or high lactate, influences the circadian gating of their function."

      4) On line 119-122, authors describe a method for polarization of macrophages. They then reference one gene to confirm each macrophage polarization state. To more definitively corroborate proper macrophage polarization, authors should perform qPCR for additional target genes that are associated with each phenotype. For example, Socs3, CD68, or CD80 for M1, and CD163 or VEGF for M2. Alternatively, the authors should cite previous literature validating this in vitro polarization model.

      We appreciate Reviewer #3's suggestion to better the phenotypic identity of our polarization models with additional canonical markers. To address this point, we have expanded our panel using transcriptional markers commonly used in the murine polarization model for M1 macrophages such as Tnfa, Il6, and Il1b. As discussed in the response to Reviewer #1's minor point 5 and Reviewer #2's major point 2, we have also expanded our panel to include additional markers for M2 such as Vegf, Retnla, Ym1, Mgl1, and CD206. We have added these new data to Supplementary Figure 1. Finally, we have added additional citations for the in vitro polarization models. Please see the modified text and new data, below.

      "As macrophages are a phenotypically heterogeneous population in the TME, we first sought to understand whether diversity in macrophage phenotype could translate to diversity in circadian rhythms of macrophages. To this end, we used two well-established in vitro polarization models to study distinct macrophage phenotypes[5, 60-63]. For a model of pro-inflammatory macrophages, we stimulated macrophages with IFNγ (interferon γ) and LPS (lipopolysaccharide) to elicit a pro-inflammatory phenotype[60, 64]. These macrophages are often referred to as 'M1' and are broadly viewed as anti-tumorigenic, and we will refer to them throughout this paper as pro-inflammatory macrophages[65, 66]. For a model at the opposite end of the phenotypic spectrum, we stimulated macrophages with IL-4 and IL-13[60, 67]. While these type 2 stimuli play a role in the response to parasites and allergy, they are also major drivers of wound healing; in line with this, IL-4 and IL-13-stimulated macrophages have been well-characterized to adopt gene expression profiles associated with wound-healing and anti-inflammatory macrophage phenotypes[68-71]. As such, these macrophages are often used as a model to study pro-tumorigenic macrophages in vitro and are often referred to as 'M2' macrophages; throughout this paper, we will refer to IL-4 and IL-13-stimulated macrophages as pro-resolution macrophages[66, 72, 73]. Consistent with previous studies, we found that genes associated with anti-inflammatory and pro-resolution programming characteristic of IL-4 and IL-13-stimulated macrophages such as Arg1, Retnla, Chil3 (Ym1), Clec10a (MGL1), and Mrc1 (CD206) were induced in IL-4 and IL-13-stimulated macrophages, but not IFNγ and LPS-stimulated macrophages. In contrast, genes associated with pro-inflammatory activity characteristic of IFNγ and LPS-stimulated macrophages such as Nos2 (iNOS), Tnfa, Il1b, and Il6 were induced in IFNγ and LPS-stimulated macrophages, but not IL-4 and IL-13-stimulated macrophages (Supplementary Figure 1)[28, 30, 65, 71, 74, 75]. This indicates that macrophages stimulated with IL-4 and IL-13 were polarized toward a pro-resolution phenotype, while macrophages stimulated with IFNγ and LPS were polarized toward a pro-inflammatory phenotype.

      5) Several portions of the manuscript are unnecessarily long, including the intro and discussion. Please consolidate the text. The results section is also very lengthy, please consider consolidation.

      We appreciate Reviewer #3's preference for a shorter manuscript. The revised manuscript, in response to the many Reviewer comments and requests, contains many new pieces of data, and we have taken care to describe these new data as briefly and simply as possible. In preparation for this Revision, we also removed and shortened several sections of the Results and Discussion where we felt extra explanation was not necessary. We will work with the editor of the journal we submit to ensure the length of the manuscript sections is compliant with the journal's guidelines.

      6) The authors find that macrophage phenotype impacts rhythmicity. However, there is no mechanistic understanding of why this occurs. The authors should provide some mechanistic insight on this topic in the discussion.

      We agree with Reviewer #3 that while the mechanism by which macrophage phenotype alters rhythms remains unknown, this is an important topic of discussion. While there is some literature on how circadian rhythms modulate inflammatory response (and hints at how it may influence phenotype) in macrophages, there is very little on the converse: how phenotype may influence circadian rhythms. We address this point by expanding on our Discussion - please see the modified text below.

      "Elucidating the role of circadian rhythms in regulation of macrophage biology necessitates a better understanding of the crosstalk between phenotype and circadian rhythms. Although lactate polarizes macrophages toward a pro-resolution phenotype similar to acidic pH[30, 93], exposure to lactate had different effects on circadian rhythms - and in some cases, circadian clock gene expression - than exposure to acidic pH (Figure 4). Sensing of lactate occurs through different pathways than acid-sensing, which may contribute to the different ways in which these two stimuli modulate circadian rhythms of macrophages[111]. One previously published finding that may offer mechanistic insight into how phenotype can influence circadian rhythms is the suppression of Bmal1 by LPS-inducible miR-155[54]. It has also been observed that RORα-mediated activation of Bmal1 transcription is enhanced by PPARγ co-activation[112]. In macrophages, PPARγ expression is induced upon stimulation with IL-4 and plays a key role in alternative activation of macrophages, promoting a pro-resolution macrophage phenotype, and supporting resolution of inflammation[113-115]. Such observations prompt the question of whether there are yet-unidentified factors induced downstream of various polarizing stimuli that can modulate expression of circadian genes at the transcriptional and protein levels. Further work is required to understand the interplay between macrophage phenotype and circadian rhythms."

      7) The data presented in Figure 9 is very intriguing and arguably the strongest aspect of the paper. To strengthen the point, the authors could repeat this experiment with an additional cell model, another PDAC line or a different cancer line.

      We appreciate Reviewer #3's comment about the impact of tumor growth data. Indeed, our finding that deletion of Bmal1 in co-injected macrophages accelerated PDAC growth has been recapitulate by others in different cancer models. This lends strength to our observations. We discuss and cite complementary work on macrophage rhythms and tumor growth in other models of cancer the Discussion, please see below.

      "Data from our lab and others suggest that disruption of the macrophage-intrinsic circadian clock accelerates tumor growth, indicating that circadian regulation of macrophages is tumor-suppressive in models of PDAC (our work) and melanoma [109]. This agrees with complementary findings that behavioral disruption of circadian rhythms in mice (through chronic jetlag) disrupts tumor macrophage circadian rhythms and accelerates tumor growth[56]."

      Minor Comments:

      1) Data is Figure 2 is interesting and the impact on circadian rhythms is clear based on changes in amplitude and period. However, though the impact on period and amplitude is clear from Figures 2A-C, the changes in circadian gene expression are less clear. For instance, though amplitude is up in 2B, amplitude is suppressed in 2C. However, that does not appear to be reflected in the gene expression data in Figures 2E and F. The authors should comment on this.

      Reviewer #3 correctly points out that there appear to be discrepancies between the LumiCycle data in Figure 2 and the circadian gene expression data in Figure 2. This discrepancy is perhaps unsurprising given that the gene expression data is only a short time course over 12 hours, while the LumiCycle data are collected over a course of 3 days. The gene expression data do not allow us to determine changes in period or rhythm. Another point of interest is that it's been shown that circadian regulation occurs on many different levels (transcriptional, post-transcriptional, translational, post-translational). As result of this, circadian patterns observed in gene transcripts don't always match those of their encoded proteins; just the same, circadian patterns of proteins aren't always reflected in their encoding gene transcripts (Collins 2021 Genome Res). Due to this multi-level regulation, we propose that the results of the LumiCycle analysis, which measures PER2-Luc levels, are a more robust readout of rhythms because they are further downstream of the molecular clock than transcriptional readouts. That said, observing changes at both the protein (by Lumicycle) and transcriptional level confirm that all components of the clock are altered by acidic pH, even if the way in which they are altered appears to differ. We have incorporated the points we raised above into the Results section.

      Please see the modified text below.

      "Low pH was also observed to alter the expression of the circadian clock genes Per2, Cry1, and Nr1d1 (REV-ERBα) over time across different macrophage phenotypes, confirming that multiple components of the circadian clock are altered by acidic pH (Figure 2D-F). Notably, the patterns in expression of circadian genes did not always match the patterns of PER2-Luc levels observed by LumiCycle. This is perhaps unsurprising, as circadian rhythms are regulated at multiple levels (transcriptional, post-transcriptional, translational, post-translational); as a result, circadian patterns observed in circadian proteins such as PER2-Luc do not always match those of their gene transcripts[77]."

      2) On line 156-158, authors describe damping rate. I believe the authors are trying to say that damping rate increases as the time it takes cells to desynchronize decreases and vice versa. However, this point needs to be better explained.

      We thank Reviewer #3 for bringing to our attention that this was not communicated clearly in the text. We have adjusted our explanation to be clearer. Please see the modified text below.

      "Damping of rhythms in most free-running cell populations (defined as populations cultured in the absence of external synchronizing stimuli) occurs naturally as the circadian clocks of individual cells in the population become desynchronized from each other; thus, damping can be indicative of desynchrony within a population[84]. The damping rate increases as the time it takes for rhythms to dissipate decreases; conversely, as damping rate decreases as the time it takes for rhythms to dissipate increases."

      3) Data presented in Figures 3 and 4 are different in terms of the impact of changing the pH. The source of the macrophages is different, but the authors could clarify this further.

      We thank Reviewer #3 for this comment. Our conclusion is that the impact of low pH is largely similar in Figure 3 (peritoneal macrophages) and Figure 4 (BMDMs exposed to low pH and lactate). In both Figures 3 and 4, exposure to acidic pH by culturing macrophages at pH 6.5 increased amplitude, decreased period, and increased damping rate compared to macrophages cultured at neutral pH 7.4.

      4) For heatmaps shown in Figures 7 and 8, please calculate covariance and display asterisks where P We thank Reviewer #3 for the excellent suggestion to use an additional approach to asses circadian clock status in samples by measuring co-variance in the circadian clock gene network. To address this point, we have performed weighted gene co-expression network analysis (WGCNA) to calculate covariance, as was originally performed in Chun and Fortin et al Science Advances 2022. For the samples analyzed in Figure 7 (now Figure 6), we have added these data to the figure. We have applied this analysis to a new set of human data that we analyzed and added it to the new Figure 7. Finally, for the samples analyzed in Figure 8, we have added these data as a new Supplementary Figure 9. Please see the data and modified text below.

      Figure 6

      "Weighted gene co-expression network analysis (WGCNA) has been used as an alternate approach to measure the co-variance between clock genes and thus assess bi-directional correlations among the core clock gene network in healthy tissue and tumor samples [103]. In line with the circadian disorder observed by CCD, while many bi-directional correlations among the core clock gene network were significant and apparent in wild type peritoneal macrophages, these relationships were altered or abolished within BMAL1 KO peritoneal macrophages and TAM samples, and in some cases replaced by new relationships (Figure 6E). This indicates that there is population-level disorder in the circadian rhythms of tumor-associated macrophages in murine lung cancer."

      Figure 7

      "We next assessed the status of the circadian clock in human TAMs from NSCLC patients. We performed CCD with publicly available RNA-seq data of tumor-adjacent macrophages and tumor-associated macrophages from NSCLC patients, using alveolar macrophages from healthy donors as a control[104, 105]. To assess the contribution of the acidic TME to circadian disorder, we subset TAM NSCLC patient samples into groups (Crem high TAMs and Crem low TAMs) based on median Crem expression. Notably, in macrophages from human NSCLC there was a trend toward disorder in Crem low but not Crem high TAM samples (Figure 7A,B). Additionally, the co-variance among core clock genes observed in alveolar macrophages from healthy donors was absent within Crem low and Crem high TAM samples (Figure 7C). In all, these data indicate that there is population-level disorder in the circadian rhythms of tumor-associated macrophages in humans and mice, suggesting that circadian rhythms are indeed altered in macrophages within the TME."

      Supplementary Figure 9

      "CCD score worsened as populations became increasingly desynchronized, with the 12hr desynchronized population having a significantly worse CCD score than synchronized, homogenous macrophage population (Figure 8C). This indicates that as circadian rhythms of individual macrophages within a population become more different from each other, circadian disorder increases at the population-level. This is further supported by WGCNA, which revealed that the significant co-variance of circadian clock genes in the synchronized population was progressively altered and lost as the population is increasing desynchronized to 12 hours (Supplementary Figure 9)."

      Reviewer #3 (Significance (Required)):

      This is an important study that is well-done. It is the feeling of the reviewer that the study warrants a revision, at the discretion of the editor. The study represents an important understanding of the crosstalk between cancer cells and immune cells as well as the understanding of how the TME disrupts circadian rhythms.

      We thank Reviewer #3 for their comments regarding the impact and significance of our work. As shown by the comments above, we are confident we have fully addressed each of the points that were made to result in a much-improved revised manuscript.




    1. Author response:

      The following is the authors’ response to the original reviews.

      Response to Reviewer 1

      (Cys25)PTH(1-84) does not show efficacy surpassing that of the previously used rhPTH(1-34). This needs to be discussed biologically and clinically.

      Thank you very much for your valuable comments for enhancing the manuscript. We appreciate your input and have noted that this aspect was not addressed in the discussion. The authors have included the following paragraph in discussion section.

      “This biological difference is thought to be due to dimeric R25CPTH(1-34) exhibiting a more preferential binding affinity for the RG versus R0 PTH1R conformation, despite having a diminished affinity for either conformation. Additionally, the potency of cAMP production in cells was lower for dimeric R25CPTH compared to monomeric R25CPTH, consistent with its lower PTH1R-binding affinity.  (Noh et al., 2024) One of the potential clinical advantages of dimeric R25CPTH(1-34) is its partial agonistic effect in pharmacodynamics. This property may allow for a more fine-tuned regulation of bone metabolism, potentially reducing the risk of adverse effects associated with full agonism, such as hypercalcemia and bone resorption by osteolcast activity. Moreover, the dimeric form may offer a more sustained anabolic response, which could be beneficial in the context of long-term treatment strategies. (Noh et al., 2024) Also, the effects of dimer were prominent, as we mentioned better bone formation than the control group.” (2nd paragraph, Discussion section)

      The terms (Cys25)PTH(1-84) and Dimeric R25CPTH(1-34) are being used interchangeably and incorrectly. A unification of these terms is necessary.

      We totally agree with the reviewer’s notion. R25CPTH(1-84) represents mutated human PTH, rhPTH(1-34) and dimeric R25CPTH(1-34) are synthesized PTH analogs. To clarified the terminology, we thus have changeed the terminology in the manuscript appear in red.

      The figure legend is incorrect. Not all figures are described, and even though there are figures from A to I, only up to E is explained, or the content is different.

      We apologize for our negligence. As suggested by a reviewer, we've fixed the figure legends throughout before the list of references in the manuscript as follows.

      “Figure legends

      Figure 1. Micro-CT analysis (A-D) Experimental design for the controlled delivery of rhPTH(1-34) and dimeric R25CPTH(1-34) in ovariectomized beagle model. Representative images for injection and placement of titanium implant. (E) Micro-CT analysis. bone mineral density (BMD), bone volume (TV; mm3), trabecular number (Tb.N; 1/mm), trabecular thickness (Tb. Th; um), trabecular separation (Tb.sp; ㎛). Error bars indicate standard deviation. Data are shown as mean ± s.d. *p<0.05, **p<0.01, ***p<0.001, n.s., not significant.  P, posterior. R, right

      Figure 2. (A-I) Histological analysis of the different groups stained in Goldner’s trichrome. The presence of bone is marked by the green color and soft tissue in red. Red arrows indicate the position with soft tissues without bone around the implant threads. The area of bone formed was the widest in the rhPTH(1-34)-treated group. In the dimeric R25CPTH(1-34)treated group, there is a greater amount of bone than vehicle-treated group. Green arrows represent the bone formed over the implant. blue dotted line, margin of bone and soft tissue; Scale bars: 1mm

      Figure 3. Histological analysis using Masson trichrome staining results in the rhPTH(1-34) and dimeric R25CPTH(1-34)-treated group (A-L) Masson trichrome-stained sections of cancellous bone in the mandibular bone. The formed bone is marked by the color red. Collagen is stained blue. Black dotted box magnification region of trabecular bone in the mandible. Scale bars, A-C, G-I: 1mm; D-F, J-L: 200 ㎛

      Figure 4. Immunohistochemical analysis using TRAP staining for bone remodeling activity (A-L) TRAP staining is used to evaluate bone remodeling by staining osteoclasts. Osteoclasts is presented by the purple color. Black dotted box magnification region of trabecular bone in the mandible. (M, N) The number of TRAP-positive cells in the mandible of the rhPTH(1-34) and dimeric R25CPTH(1-34)-treated beagles. Scale bars, A-C, G-I: 1mm; D-F, J-L: 200 ㎛. Error bars indicate standard deviation. Data are shown as mean ± s.d. *p<0.05, **p<0.01, n.s., not significant

      Figure 5. Measurement of biochemical Marker Dynamics in serum. The serum levels of calcium, phosphorus, P1NP, and CTX across three time points (T0, T1, T2) following treatment with dimeric dimeric R25CPTH(1-34), rhPTH(1-34), or control. (A-B) Calcium and phosphorus levels exhibit an upward trend in response to both PTH treatments compared to control, suggesting enhanced bone mineralization. (C) P1NP levels, indicative of bone formation, remain relatively unchanged across time and treatments. (D) CTX levels, associated with bone resorption, show no significant differences between groups. Data points for the dimeric R25CPTH(1-34), rhPTH(1-34), and control are marked by squares, circles, and triangles, respectively, with error bars representing confidence intervals.

      Supplementary Figure. Three-dimensional reconstructed image of the bone surrounding the implants. Three-dimensional reconstructed images of the peri-implant bone depicting the osseointegration after different therapeutic interventions. (A) Represents the bone response to recombinant human parathyroid hormone fragment (rhPTH 1-34) treatment, showing the most robust degree of bone formation around the implant in the three groups. (B) Shows the bone response to a modified PTH fragment (dimeric R25CPTH(1-34)), indicating a similar level of bone growth and integration as seen with rhPTH(1-34), although to a slightly lesser extent. (C) Serves as the control group, demonstrating the least amount of bone formation and osseointegration. The upper panel provides a top view of the bone-implant interface, while the lower panel offers a cross-sectional view highlighting the extent of bony ingrowth and integration with the implant surface.”

      In Figure 5, although the descriptions of T0, T1, T2 are mentioned in the method section, it would be more clear if there was a timeline like in Figure 1.

      Based on the reviewer’s advice, we have indicated the timing of T0, T1, and T2 in the materials & methods section describing the serum biochemical assay, and we have shown a timeline in figure 5.

      In Figure 5, instead of having calcium, phosphorus, P1NP, CTX graphs all under Figure 5, it would be more convenient for referencing in the text to label them as Figure 5A, Figure 5B, Figure 5C, Figure 5D.

      We totally understood the reviewer’s comment. As the reviewer’s suggested, we have corrected the labeling in the text for figure 5 as follows.

      “The levels of calcium, phosphorus, CTX, and P1NP were analyzed over time using RM-ANOVA (Figure 5). There were no significant differences between the groups for calcium and phosphorus at time points T0 and T1 (Figure 5A). However, after the PTH analog was administered at T2 (Figure 5A), the levels were highest in the rhPTH(1-34) group, followed by the dimeric R25CPTH(1-34) group, and then, lowest in the control group, which was statistically significant (Figure 5B,C). (P < 0.05) The differences between the groups over time for CTX and P1NP were not statistically significant (Figure 5D, E).”

      Significance should be indicated in the figure (no asterisk present).

      As the reviewer’s comment, we put the asterisk in the figure 5.

      Addition of Figures in Text:

      Line 112: change from "figure 2" to "figure 1" / Line 115: mention "figure 1. E"

      Line 120: refer to "figure 1. E" / Line 123: change from "figure 3" to "figure 2"

      Line 128: refer to "figure 2.A-C" / Line 137: mention "figure 3"

      Line 138: refer to "figure 3. A-L" / Line 143: mention "figure 3. A-L"

      Line 144: refer to "figure 3. E,F,K,L" / Line 148: mention "figure 4"

      Line 150: refer to "figure 4 M,N" / Line 152: mention "figure 4. M,N"

      Line 155: refer to "figure 5" / Line 157: mention "figure 5"

      Line 159: refer to "figure 5" / Line 171: mention "figure 1 E"

      Line 175: refer to "figure 2 M, N"/ Line 194: mention "figure 3"

      Above all, thank you for the reviewer’s notion. We corrected detailed figure labeling in text to red color.

      Response to Reviewer 2

      First, the authors should clarify why they compared the effects of rhPTH(1-34) and of dimeric R25C2 PTH(1-34)? In most of the parameters, rhPTH(1-34) seems to be superior to dimeric R25C2 PTH(1-34). Why did the authors insist that the anabolic effects of dimer were prominent? Even though implication of dimeric R25C2 PTH(1-34) was drawn from genetic mutation studies, the authors should describe more clearly in the discussion the potential clinical benefits of the dimeric R25C2 PTH(1-34) compared to rhPTH(1-34), especially if dimeric R25C2 PTH(1-34) has just partial agonistic effect in pharmacodynamics.

      Thank you for your insightful comments and questions regarding our results between rhPTH(1-34) and dimeric R25CPTH(1-34). rhPTH(1-34) is a well-characterized therapy for osteoporosis. In this study, rhPTH(1-34) generally showed superior outcomes in most parameters tested, the dimeric R25CPTH(1-34) exhibited specific anabolic effects that are not as pronounced with rhPTH(1-34). We recognized R25CPTH(1-34) as a anabolic effector. One of the potential advantages of dimeric R25CPTH(1-34) is its partial agonistic effect in pharmacodynamics. This property may allow for a more fine-tuned regulation of bone metabolism, potentially reducing the risk of adverse effects associated with full agonism, such as hypercalcemia and bone resorption by osteolast activity. Moreover, the dimeric form may offer a more sustained anabolic response, which could be beneficial in the context of long-term treatment strategies. Also, based on our results, we notes that the effects of dimer were prominent, as we mentioned better bone formation than the control group. We appreciate your input and have noted that this aspect was not addressed in the discussion. As a result, we have included the following paragraph in discussion section.

      “This biological difference is thought to be due to dimeric R25CPTH(1-34) exhibiting a more preferential binding affinity for the RG versus R0 PTH1R conformation, despite having a diminished affinity for either conformation. Additionally, the potency of cAMP production in cells was lower for dimeric R25CPTH compared to monomeric R25CPTH, consistent with its lower PTH1R-binding affinity.  (Noh et al., 2024) One of the potential clinical advantages of dimeric R25CPTH(1-34) is its partial agonistic effect in pharmacodynamics. This property may allow for a more fine-tuned regulation of bone metabolism, potentially reducing the risk of adverse effects associated with full agonism, such as hypercalcemia and bone resorption by osteolcast activity. Moreover, the dimeric form may offer a more sustained anabolic response, which could be beneficial in the context of long-term treatment strategies. (Noh et al., 2024) Also, the effects of dimer were prominent, as we mentioned better bone formation than the control group.” (2nd paragraph, Discussion section)

      Second, please describe the intermittent and continuous application of PTH analogues. Many of the readers may misunderstand that the authors' daily injection of PTHs were actually to mimic the clinical intermittent application or continuous one. Incorporation of the author's intention for experimental design would be more helpful for readers.

      Thank you for your insightful comments regarding the need for clearer differentiation between intermittent and continuous applications of PTH analogs in this study. We appreciate your concern that the readers may not fully grasp whether our daily injection protocol was intended to mimic clinical intermittent or continuous PTH administration. To address this, we have revised the manuscript to explicitly clarify that the daily injections of rhPTH(1-34) and dimeric R25CPTH(1-34) were designed to simulate the intermittent dosing regimen commonly used in clinical practice. This regimen is known to maximize the anabolic effects on bone while minimizing potential catabolic actions associated with more frequent or continuous hormone exposure. We have added detailed explanations in the Introduction, Methods, and Discussion sections to help readers understand our experimental design and its relevance to clinical settings.

      Introduction section

      “Administration of prathyroid hormone (PTH) analogs can be categorized into two distinct protocols: intermittent and continuous. Intermittent rhPTH(1-34) therapy, typically characterized by daily injections, is clinically used to enhance bone formation and strength. This method leverages the anabolic effects of rhPTH(1-34) without significant bone resorption, which can occur with more frequent or continuous exposure. On the other hand, continuous rhPTH(1-34) exposure, often modeled in research as constant infusion, tends to accelerate bone resorption activities, potentially leading to bone loss (Silva and Bilezikian, 2015; Jilka, 2007). Understanding these differences is crucial for interpreting the therapeutic implications of rhPTH(1-34) in bone health.”

      Silva, B. C., & Bilezikian, J. P. (2015). Parathyroid hormone: anabolic and catabolic actions on the skeleton. Current Opinion in Pharmacology, 22, 41-50.

      Jilka, R. L. (2007). Molecular and cellular mechanisms of the anabolic effect of intermittent PTH. Bone, 40(6), 1434-1446.

      Materials and Methods section

      “Each animal received one injection per day, aimed at replicating the intermittent rhPTH(1-34) exposure proven beneficial for bone regeneration and overall skeletal health in clinical settings (Neer et al., 2001; Kendler et al., 2018). This regimen was chosen to investigate the potential anabolic effects of these specific PTH analogs under conditions closely resembling therapeutic use.”

      Neer, R. M., Arnaud, C. D., Zanchetta, J. R., Prince, R., Gaich, G. A., Reginster, J. Y., Hodsman, A. B., Eriksen, E. F., Ish-Shalom, S., Genant, H. K., Wang, O., and Mitlak, B. H. (2001). Effect of Parathyroid Hormone (1-34) on Fractures and Bone Mineral Density in Postmenopausal Women with Osteoporosis. The New England Journal of Medicine, 344(19), 1434-1441.

      Kendler, D. L., Marin, F., Zerbini, C. A. F., Russo, L. A., Greenspan, S. L., Zikan, V., Bagur, A., Malouf-Sierra, J., Lakatos, P., Fahrleitner-Pammer, A., Lespessailles, E., Minisola, S., Body, J. J., Geusens, P., Moricke, R., & Lopez-Romero, P. (2018). Effects of Teriparatide and Risedronate on New Fractures in Post-Menopausal Women with Severe Osteoporosis (VERO): A Multicenter, Double-Blind, Double-Dummy, Randomized Controlled Trial. The Lancet, 391(10117), 230-240.

      Discussion section

      “The use of daily injections in this study was intended to simulate intermittent PTH therapy, a well-established clinical approach for managing osteoporosis and enhancing bone regeneration. Intermittent administration of PTH, as opposed to continuous exposure, is critical for maximizing the anabolic response while minimizing the catabolic effects that are associated with higher frequency or continuous hormone levels. Our findings support the notion that even with daily administration, both rhPTH(1-34) and dimeric dimeric R25CPTH(1-34) promote bone formation and osseointegration, consistent with the outcomes expected from intermittent therapy. It’s important for future research to consider the dosage and timing of administration to further optimize the therapeutic benefits of PTH analogs (Dempster et al., 2001; Hodsman et al., 2005).”

      Dempster, D. W., Cosman, F., Kurland, E. S., Zhou, H., Nieves, J., Woelfert, L., Shane, E., Plavetic, K., Müller, R., Bilezikian, J., & Lindsay, R. (2001). Effects of Daily Treatment with Parathyroid Hormone on Bone Microarchitecture and Turnover in Patients with Osteoporosis: A Paired Biopsy Study. Journal of Bone and Mineral Research, 16(10), 1846-1853.

      Hodsman, A. B., Bauer, D. C., Dempster, D. W., Dian, L., Hanley, D. A., Harris, S. T., Kendler, D. L., McClung, M. R., Miller, P. D., Olszynski, W. P., Orwoll, E., Yuen, C. K. (2005). Parathyroid Hormone and Teriparatide for the Treatment of Osteoporosis: A Review of the Evidence and Suggested Guidelines for Its Use. Endocrine Reviews, 26(5), 688-703.

      Third, please unify the nomenclature. Ensure consistency in the nomenclature throughout the article. Unify the naming conventions for PTH analogues, such as rhPTH(1-34) vs teriparatide and (Cys25)PTH(1-84) vs R25CPTH(1-34) vs R25CPTH(1-34) vs (1-84). Choose one nomenclature for each analogue and use it consistently throughout the article.

      We totally agree with the reviewer’s notion. R25CPTH(1-84) represents mutated human PTH, rhPTH(1-34) and dimeric R25CPTH(1-34) are synthesized PTH analogs. To clarified the terminology, we thus have changed the terminology in the manuscript appear in red.

      Response to Reviewer 3

      I would recommend to rewrite the manuscript in a form that is more understandable to the readers. In fact, it appears to me that this work was originally formatted in a way that would need the Materials and Methods to precede the results. As presented (and as requested by the eLife formatting) the Materials and Methods are available only at the end of the reading and, as a consequence, the readers needs to refer to the Materials and Methods to have a general and initial understanding of the study design (i.e. type of treatment for each group, etc are not well specified in the Results section).

      Thank you for you constructive comments and suggestions regarding the manuscript. We appreciate your feedback on the organization of the manuscript entirely. As reviewer mentioned, Materials and methods were placed after the discussion section in accordance with the format of the elife journal. For a better and initial understanding, a description of each experimental group has been added to the Results section as follow. Thank you again for your valuable comments.

      “To investigate evaluating and comparing the efficacy of rhPTH(1-34) and the dimeric R25CPTH(1-34) in promoting bone regeneration and healing in a clinically relevant animal model. In our study, beagle dogs were selected as the model due to their anatomical similarity to human oral structures, suitable size for surgeries, human-like bone turnover rates, and established oral health profiles, ensuring comparable and ethically sound research outcomes. The normal saline injected-control group, injected with 40ug/day PTH (Forsteo, Eli Lilly) group, and 40ug/day PTH analog-injected group. Animals in each group were injected subcutaneously for 10 weeks.”

    1. why the writer or speaker communicating.

      I think at times there can be more than one purpose in communication. For example, news stations are know to be informative but they do fall into those persuade, analyze, theorize and entertain components too. Adding on to what Sofia said, I think it can be beneficial to show other perspectives especially with topics that are less technical their tends to be a lot of bias. I think it's just important to distinguish the different types of communication or media which may be tricky.

    1. When it comes to history, we think in straight lines. When we imagine the progress of the next 30 years, we look back to the progress of the previous 30 as an indicator of how much will likely happen. When we think about the extent to which the world will change in the 21st century, we just take the 20th century progress and add it to the year 2000. This was the same mistake our 1750 guy made when he got someone from 1500 and expected to blow his mind as much as his own was blown going the same distance ahead. It’s most intuitive for us to think linearly, when we should be thinking exponentially. If someone is being more clever about it, they might predict the advances of the next 30 years not by looking at the previous 30 years, but by taking the current rate of progress and judging based on that. They’d be more accurate, but still way off. In order to think about the future correctly, you need to imagine things moving at a much faster rate than they’re moving now.

      In order to think about the future correctly, you need to imagine things moving at a much faster rate than they're moving now.

    1. Today, data is abundant, but for the most part, unusable. Seventy percent of a data scientist’s job is just cleansing data. The modern software architecture encourages data to be hoarded only accessible through proprietary APIs. And, even with proprietary APIs the market for data integrations is expected to grow to a trillion dollars by the end of the decade. When humanity is spending the GDP of Indonesia just so that the data in System X can work with the data in System Y, the field of software engineering has failed us. So much data - data that could be used by new startups and nonprofits that couldn’t exist today - goes unused because it’s so difficult to access.
    1. First, the complexity of modern federal criminal law, codified in several thousand sections of the United States Code and the virtually infinite variety of factual circumstances that might trigger an investigation into a possible violation of the law, make it difficult for anyone to know, in advance, just when a particular set of statements might later appear (to a prosecutor) to be relevant to some such investigation.

      If the federal government had access to every email you’ve ever written and every phone call you’ve ever made, it’s almost certain that they could find something you’ve done which violates a provision in the 27,000 pages of federal statues or 10,000 administrative regulations. You probably do have something to hide, you just don’t know it yet.

    1. Overall Rating (⭐⭐⭐☆☆): This manuscript presents a post-hoc analysis of a dietary intervention examining the short and long-term effects of a Mediterranean diet (MedDiet) versus a control diet on gestational diabetes mellitus (GDM) and metabolic syndrome (MetSyn) in overweight pregnant women. The study's focus on longer-term outcomes (3 years postpartum) is relevant and important to help understand the potential long-term effects of dietary interventions. Overall, this is a valuable contribution to the field and generally well-written. I have some suggestions for the methods and for the presentation and interpretation of the results to improve impact and clarity.

      Impact (⭐⭐⭐⭐☆): The topic of this study is important, as GDM is highly prevalent and associated with not only immediate pregnancy complications but also long-term adverse health consequences for women and their offspring. Evaluating the role of diet as a modifiable risk factor could inform recommendations for prevention. A strength is also the focus on longer-term outcomes, with results for 3 years postpartum suggesting potential long-term benefits of the MedDiet intervention: not only on measures of body weight and glucose management, but also on dietary habits, as women who were part of the intervention group during pregnancy, still seemed to have a better adherence to the MedDiet at follow-up.

      To enhance impact: it would be informative to read more about how this post-hoc analysis builds upon and differs from their previously reported findings on GDM reduction. Also, I would add some information on the rationale for choosing the Mediterranean diet as opposed to other healthy dietary patterns, particularly in the context of previous research focusing on carbohydrate quality and quantity.

      Methods (⭐⭐⭐☆☆): - The study is a post-hoc analysis combining three studies among pregnant women. The three studies and their differences need some more clarification. Maybe they can be summarized in a figure or table? Also, it seems that only studies 1 and 3 were randomized and controlled, while study 2 was not. I would recommend to do a sensitivity analysis including only studies 1 and 3. - The specific dietary recommendations provided to both the intervention and control groups need more detail. It was not completely clear to me whether the Mediterranean diet group received dietary recommendations beyond increasing olive oil and nut consumption. Similarly, it's not clear if the control group received any standard dietary recommendations, beyond limiting olive oil and nut intake. - Information on the number of women who did not participate in the follow-up measurements or who were excluded by the researchers (e.g. with a BMI<25) should be provided for each study and study group, along with reasons if known. A flow diagram would be helpful to show how many women were included in the original studies, how many participated in the follow-up, and reasons for loss to follow-up or exclusion from the current analysis. - Remove p-values for differences in baseline values between groups in Table 1, as these can be misleading and are not useful after randomization. Although this is still often included in studies, this practice has been discouraged, also in e.g. the CONSORT statement (e.g. doi: 10.1016/j.jclinepi.2010.03.004 or doi: 10.1016/S0140-6736(00)02039-0). - Is any data on compliance available? - Minor: Write out abbreviations such as IG and CG in full the first time they appear in the main text, not just in the abstract. - Minor: In section 2.4 (analysis), I would suggest to remove the word 'adverse' when describing the outcomes assessed, as all tests were indicated to be 2-sided (so not testing only adverse effects but in either direction).

      Results (⭐⭐⭐☆☆): - I would consider removing the comparison of the GDM versus the GTN group from Table 4 and the results section. This additional set of analyses does not provide evidence on the intervention and, in my opinion, does not add substantially to the main analyses and may be confusing (especially when presented in the same table as intervention results). - Minor: Please check that all variables in Table 1 include units (e.g., insulin, TG).

      Discussion(☆☆☆☆☆): <br /> - In the conclusions, I would avoid emphasizing a lower risk of MetSyn at 3 years postpartum. Instead, focus on the individual components that showed significant differences (glucose regulation, BMI/Waist circumference), as the effect on the composite score for MetSyn seems to be driven by these components only. - Although loss to follow-up is discussed in the context of sample size and statistical power, I would like to also see information and discussion on whether this loss to follow-up may be selective, how this might differ across groups. - Minor: suggestion to rephrase the last sentence of the discussion regarding the use of validated questionnaires. While questionnaires are useful and acceptable, they still have limitations such as recall bias and should not be presented as overcoming all limitations or providing a fully objective quantification. - I think the observed long-term effects on dietary behavior are interesting and deserve more attention in for example the discussion section.

      Reviewer Information Dr. Trudy Voortman has a PhD in nutritional epidemiology, 15+ years of research experience, focusing on the role of nutrition and lifestyle in population health across the lifecourse.

      Dr. Trudy Voortman on ResearchHub: https://www.researchhub.com/user/1791011/overview

      ResearchHub Peer Reviewer Statement: This peer review has been uploaded from ResearchHub as part of a paid peer review initiative. ResearchHub aims to accelerate the pace of scientific research using novel incentive structures.

    1. How to Celebrate Your Wins at Work Without Coming Across as a Jerk

      By Jessica Chen

      • Introductory Insight
      • One of the best ways to advance your career is to have your efforts recognized, yet many deflect compliments and minimize their contributions.<br /> “Isn’t it ironic that one of the best ways to accelerate our career is to have people see and recognize your efforts, yet for many of us, when that happens, such as when we get complimented or praised by our team, we instantly deflect and minimize the contribution?”

      • Challenges in Self-Promotion

      • Some find it easy to share their thoughts and highlight their work, while others find it challenging due to cultural teachings of modesty and humility.<br /> “For some of us, sharing what’s on our mind and highlighting our work comes easy... But for others, the idea of putting ourselves out there...feels challenging.”
      • It’s not just about being introverted or extroverted but also about cultural upbringing that discourages self-promotion.<br /> “For some of us, talking about ourselves wasn’t what we were taught to do... Instead, we were taught to minimize the spotlight and focus on getting the work done.”

      • Necessity of Celebrating Wins

      • To progress in your career, it’s essential to both do good work and confidently talk about your impact.<br /> “We need to get things done and confidently talk about our impact because when we do, we highlight our genius and keep ourselves top of mind for bigger opportunities at work.”
      • Celebrating wins is not optional but a necessary part of professional growth.<br /> “Celebrating our wins isn’t a nice to do, it’s a must do.”

      • Reframing Misconceptions

      • Misconception 1: Celebrating wins is selfish.
        • Reframe: It’s part of your job to communicate your work and its impact.<br /> “We celebrate our wins because it’s part of the work we do.”
      • Misconception 2: Celebrating wins will annoy others.
        • Reframe: Use tact and emphasize the benefit for the team.<br /> “We can do this by considering our tone of voice and structuring our message so it’s leading with the benefit for the team and how it has helped them.”
      • Misconception 3: Celebrating wins is complex.

        • Reframe: Simple gestures, like forwarding a client’s compliment, can be effective.<br /> “Sometimes the most effective way to highlight our wins is to approach it in the simplest of ways.”
      • Communication Strategies: ABC Checklist

      • A – Articulate the Benefit: Explain how your achievements help others.<br /> “How did your accomplishments help others?”
      • B – Be Open About the Process: Share the steps taken to accomplish the task.<br /> “What steps did you take to accomplish this task?”
      • C – Communicate Using Power Words: Use emotionally impactful words to convey your enthusiasm.<br /> “What emotions did you feel with this win? Use words like excited, happy, proud.”

      • Practical Tips

      • Create a "Yay Folder" in your email to store positive feedback for easy reference.<br /> “Create what I call a ‘Yay Folder’ in your inbox...if you ever need evidence to prove you are doing great work...you now have it stored in one place.”
      • Don't overcomplicate sharing your wins; simple expressions of excitement can be enough.<br /> “We shouldn’t overthink how we celebrate our wins at work.”

      • Final Thoughts

      • Celebrating your achievements helps reinforce the value of your work and builds your professional reputation.<br /> “Celebrating your wins is knowing your work, effort, and impact matter.”
      • Be your own cheerleader to ensure recognition of your accomplishments.<br /> “If you’re not your own best cheerleader, who will be?”

      This summary encapsulates the main ideas and actionable advice provided by Jessica Chen on effectively celebrating your wins at work without coming off as arrogant.

    1. Reviewer #2 (Public Review):

      Li et al present a method to extract "behaviorally relevant" signals from neural activity. The method is meant to solve a problem which likely has high utility for neuroscience researchers. There are numerous existing methods to achieve this goal some of which the authors compare their method to-thankfully, the revised version includes one of the major previous omissions (TNDM). However, I still believe that d-VAE is a promising approach that has its own advantages. Still, I have issues with the paper as-is. The authors have made relatively few modifications to the text based on my previous comments, and the responses have largely just dismissed my feedback and restated claims from the paper. Nearly all of my previous comments remain relevant for this revised manuscript. As such, they have done little to assuage my concerns, the most important of which I will restate here using the labels/notation (Q1, Q2, etc) from the reviewer response.

      (Q1) I still remain unconvinced that the core findings of the paper are "unexpected". In the response to my previous Specific Comment #1, they say "We use the term 'unexpected' due to the disparity between our findings and the prior understanding concerning neural encoding and decoding." However, they provide no citations or grounding for why they make those claims. What prior understanding makes it unexpected that encoding is more complex than decoding given the entropy, sparseness, and high dimensionality of neural signals (the "encoding") compared to the smoothness and low dimensionality of typical behavioural signals (the "decoding")?

      (Q2) I still take issue with the premise that signals in the brain are "irrelevant" simply because they do not correlate with a fixed temporal lag with a particular behavioural feature hand-chosen by the experimenter. In the response to my previous review, the authors say "we employ terms like 'behaviorally-relevant' and 'behaviorally-irrelevant' only regarding behavioral variables of interest measured within a given task, such as arm kinematics during a motor control task.". This is just a restatement of their definition, not a response to my concern, and does not address my concern that the method requires a fixed temporal lag and continual decoding/encoding. My example of reward signals remains. There is a huge body of literature dating back to the 70s on the linear relationships between neural and activity and arm kinematics; in a sense, the authors have chosen the "variable of interest" that proves their point. This all ties back to the previous comment: this is mostly expected, not unexpected, when relating apparently-stochastic, discrete action potential events to smoothly varying limb kinematics.

      (Q5) The authors seem to have missed the spirit of my critique: to say "linear readout is performed in motor cortex" is an over-interpretation of what their model can show.

      (Q7) Agreeing with my critique is not sufficient; please provide the data or simulations that provides the context for the reference in the fano factor. I believe my critique is still valid.

      (Q8) Thank you for comparing to TNDM, it's a useful benchmark.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Reviewer #1 (Recommendations For The Authors):

      Although the manuscript is well organized and written, it could be largely improved and therefore made more plausible and easier to read. See my point-by-point comments listed below:

      (1) The introduction section is a bit overloaded with some unnecessary information. For example, the authors discussed the relationship between neurotransmitters in the prefrontal and striatum and substance use/sustained attention. However, the results are related to neither the neurotransmitters nor the striatum. In addition, there is a contradictory description about neurotransmitters there, Nicotine/THC leads to increased neurotransmitters, and decreased neurotransmitters is related to poor sustained attention. Does that mean that the use of Nicotine/THC could increase sustained attention?

      Thanks for this insightful question. We understand your concern regarding the seemingly contradictory statements about neurotransmitters and sustained attention. Previous studies have shown that acute administration of nicotine can improve sustained attention (Lawrence et al., 2002; Potter and Newhouse, 2008; Valentine and Sofuoglu, 2018; Young et al., 2004). On the other hand, the acute effects of smoking cannabis on sustained attention are mixed and depend on factors such as dosage and individual differences (Crean et al., 2011). For instance, a previous study (Hart et al., 2001) found that performance on a tracking task, which requires sustained attention, was found to improve significantly after smoking cannabis with a high dose of THC, albeit in experienced cannabis users. However, chronic substance use, including nicotine and cannabis, has been associated with impaired sustained attention (Chamberlain et al., 2012; Dougherty et al., 2013).

      To address your concerns and improve clarity and succinctness of the Introduction, we have removed the description of neurotransmitters from the Introduction. This revision should make the introduction more concise and focus on the direct relationships pertinent to our study.

      (2) It is a bit hard to follow the story for the readers because the Results section went straight into detail. For example, the authors directly introduced that they used the ICV from the Go trials to index sustained attention without basic knowledge about the task. Why use the ICV of Go trials instead of other trials (i.e., successful stop trials) as an index of sustained attention? I suggest presenting the subjects and task details about the data before the detailed behavioral results. The results section should include enough information to understand the presenting results for the readers, rather than forcing the reader to find the answer in the later Methods section.

      We appreciate your suggestion to provide more context about the task and ICV before diving into the detailed behavioural results.

      We used the ICV derived from the Go trials instead of Success stop trials as an index of sustained attention, based on the nature of the stop-signal task and the specific data it generates. Previous studies have indicated that reaction time (RT) variability is a straightforward measure of sustained attention, with increasing variability thought to reflect poorer ability to sustain attention (Esterman and Rothlein, 2019). RT variability is defined as ICV, calculated as the standard deviation of mean Go RT divided by the mean Go RT from Go trials (O'Halloran et al., 2018). The stop signal task includes both Go trials and stop trials. During Go trials, participants are required to respond as quickly and accurately as possible to a Go signal, allowing for the recording of RT for calculating ICV. In contrast, stop trials are designed to measure inhibitory control, where successful response inhibition results in no RT or response recorded in the output. Therefore, Go trials are specifically used to assess sustained attention, while Stop trials primarily assess inhibitory control (Verbruggen et al., 2019).

      We acknowledge the importance of providing this contextual information within the Results section to enhance reader understanding. We have added this information before presenting the behavioural results on Page 6.

      Results

      (1) Behavioural changes over time

      Reaction time (RT) variability is a straightforward measure of sustained attention, with increasing variability thought to reflect poor sustained attention. RT variability is defined as intra-individual coefficient of variation (ICV), calculated as the standard deviation of mean Go RT divided by the mean Go RT from Go trials in the stop signal task. Lower ICV indicates better sustained attention.

      (3) The same problem for section 2 in the Results. What are the predictive networks? Are the predictive networks the same as the networks constructed based on the correlation with ICV? My intuitive feeling is that they are the circular analyses here. The positive/negative/combined networks are calculated based on the correlation between the edges and ICV. Then the author used the network to predict the ICV again. The manipulation from the raw networks (I think they are based on PPI) to the predictive network, and the calculation of the predicted ICV are all missing. The direct exposure of the results to the readers without enough detailed knowledge made everything hard to digest.

      We thank the Reviewer for the insightful comment. We agree with the need for more clarity regarding the predictive networks and the CPM analysis before presenting results. CPM, a data-driven neuroscience approach, is applied to predict individual behaviour from brain functional connectivity (Rosenberg et al., 2016; Shen et al., 2017). The CPM analysis used the strength of the predictive network to predict the individual difference in traits and behaviours. CPM includes several steps: feature selection, feature summarization, model building, and assessment of prediction significance (see Fig. S1).

      During feature selection, we assessed whether connections between brain areas (i.e., edges) in a task-related functional connectivity matrix (derived from general psychophysiological interaction analysis) were positively or negatively correlated with ICV using a significance threshold of P < 0.01. These positively or negatively correlated connections are regarded as positive or negative network, respectively. The network strength of the positive network (or negative network) was determined in each individual by summing the connection strength of each positively (or negatively) correlated edge. The combined network was determined by subtracting the strength of the negative network from the positive network. Next, CPM built a linear model between the network strength of the predictive network and ICV. This model was initially developed using the training set. The predictive networks were then applied to the test set, where network strength was calculated again, and the linear model was used to predict ICV using k-fold cross-validation. Following your advice, we have updated it in the Results section to include these details on Page 7.

      Results

      (2) Cross-sectional brain connectivity

      This study employed CPM, a data-driven neuroscience approach, to identify three predictive networks— positive, negative, and combined— that predict ICV from brain functional connectivity. CPM typically uses the strength of the predictive networks to predict individual differences in traits and behaviors. The predictive networks were obtained based on connectivity analyses of the whole brain. Specifically, we assessed whether connections between brain areas (i.e., edges) in a task-related functional connectivity matrix derived from generalized psychophysiological interaction analysis were positively or negatively correlated with ICV using a significance threshold of P < 0.01. These positively or negatively correlated connections were regarded as positive or negative network, respectively. The network strength of positive networks (or negative networks) was determined for each individual by summing the connection strength of each positively (or negatively) correlated edge. The combined network was determined by subtracting the strength of the negative network from the positive network. We then built a linear model between network strength and ICV in the training set and applied these predictive networks to yield network strength and a linear model in the test set to calculate predicted ICV using k-fold cross validation.

      (4) The authors showed the positive/negative/combined networks from both Go trials and successful stop trials can predict the ICV. I am wondering how the author could validate the specificity of the prediction of these positive/negative/combined networks. For example, how about the networks from the failed stop trials?

      We appreciate the opportunity to clarify the specificity of the predictive networks identified in our study. Here is a more detailed explanation of our findings and their implications.

      To validate the specificity of the sustained attention network identified from CPM analysis, we calculated correlations between the network strength of positive and negative networks and performances from a neuropsychology battery (CANTAB) at each timepoint separately. CANTAB includes several tasks that measure various cognitive functions, such as sustained attention, inhibitory control, impulsivity, and working memory. We found that all positive and negative networks derived from Go and Successful stop trials significantly correlated with a behavioural assay of sustained attention – the rapid visual information processing (RVP) task – at ages 14 and 19 (all P values < 0.028). Age 23 had no RVP task data in the IMAGEN study. There were sporadic significant correlations between constructs such as delay aversion/impulsivity and negative network strength, for example, but the correlations with the RVP were always significant. This demonstrates that the strength of the sustained attention brain network was specifically and robustly correlated with a typical sustained attention task, rather than other cognitive measures. The results are described in the main text on Page 8 and shown in Supplementary materials (Pages 1 and 3) and Table S12.

      In addition, we conducted a CPM analysis to predict ICV using gPPI under Failed stop trials. Our findings showed that positive, negative, and combined networks derived from Failed stop trials significantly predicted ICV: at age 14 (r = 0.10, P = 0.033; r = 0.19, P < 0.001; and r = 0.17, P < 0.001, respectively), at age 19 (r = 0.21; r = 0.18; and r = 0.21, all P < 0.001, respectively), and at age 23 (r = 0.33, r = 0.35, and r = 0.36, respectively, all P < 0.001). Similar results were obtained using a 5-fold CV and leave-site-out CV.

      Our analysis further showed that task-related functional connectivity derived from Go trials, Successful Stop trials, and Failed Stop trials could predict sustained attention across three timepoints. However, the predictive performances of networks derived from Go trials were higher than those from Successful Stop and Failed Stop trials. This suggests that sustained attention is particularly crucial during Go trials when participants need to respond to the Go signal. In contrast, although Successful Stop and Failed Stop trials also require sustained attention, these tasks primarily involve inhibitory control along with sustained attention.

      Taken together, these findings underscore the specificity of the predictive networks of sustained attention. We have updated these results in the Supplementary Materials (Pages 3-5 and Page 7 ):

      Method

      CPM analysis using Failed stop trials

      We performed another CPM analysis using Failed stop trials using gPPI matrix obtained from the second GLM, described in the main text. The CPM analysis was conducted using 10-fold CV, 5-fold CV and leave-site-out CV.

      Results

      CPM predictive performance under Failed stop trials

      Positive, negative, and combined networks derived from Failed stop trials significantly predicted ICV: at age 14 (r = 0.10, P = 0.033; r = 0.19, P < 0.001; and r = 0.17, P < 0.001, respectively), at age 19 (r = 0.21; r = 0.18; and r = 0.21, all P < 0.001, respectively), and at age 23 (r = 0.33, r = 0.35, and r = 0.36, respectively, all P < 0.001). We obtained similar results using a 5-fold CV and leave-site-out CV (Table S6).

      Discussion

      Specificity of the prediction of predictive networks

      We found that task-related function connectivity derived from Go trials, Successful stop trials, and Failed stop trials successfully predicted sustained attention across three timepoints. However, predictive performances of predictive networks derived from Go trials were higher than those derived from Successful stop trials and Failed stop trials. These results suggest that sustained attention is particularly crucial during Go trials when participants need to respond to the Go signal. In contrast, although Successful Stop and Failed Stop trials also require sustained attention, these tasks primarily involve inhibitory control along with sustained attention.

      (5) The author used PPI to define the connectivity of the network. I am not sure why the author used two GLMs for the PPI analysis separately. In the second GLM, Go trials were treated as an implicit baseline. What does this exactly mean? And the gPPI analysis across the entire brain using the Shen atlas is not clear. Normally, as I understand, the PPI/gPPI is conducted to test the task-modulated connectivity between one seed region and the voxels of the whole rest brain. Did the author perform the PPI for each ROI from Shen atlas? More details about how to use PPI to construct the network are required.

      Thank you for your insightful questions. Here, we’d like to clarify how we applied generalized PPI across the whole brain using the Shen atlas and why we used two separate GLMs for the gPPI analysis.

      Yes, PPI is conducted to test the task-modulated connectivity between one seed region and other brain areas. This method can be both voxel-based and ROI-based. In our study, we performed ROI-based gPPI analysis using Shen atlas with 268 regions. Specifically, we performed the PPI on each seed region of interest (ROI) to estimate the task-related FC between this ROI and the remaining ROI (267 regions) under a specific task condition. By performing this analysis across each ROI in the Shen atlas, we generated a 268 × 268 gPPI matrix for each task condition. The matrices were then transposed and averaged with the original matrices, which yielded symmetrical matrices, which were subsequently used for CPM analysis.

      Regarding the use of two separate GLMs for the gPPI analysis, our study aimed to define the task-related FC under two conditions: Go trials and Successful stop trials. The first GLM including Go trials was built to estimate the gPPI during Go trials. However, due to the high frequency of Go trials in the stop signal task, it is common to regard the Go trials as an implicit baseline, as in previous IMAGEN studies (D'Alberto et al., 2018; Whelan et al., 2012). Therefore, to achieve a more accurate estimation of FC during Successful stop trials, we built a second GLM specifically for these trials. Accordingly, we have updated it in the Method Section in the main text on Page 16.

      Method

      2.5 Generalized psychophysiological interaction (gPPI) analysis

      In this study, we adopted gPPI analysis to generate task-related FC matrices and applied CPM analysis to investigate predictive brain networks from adolescents to young adults. PPI analysis describes task-dependent FC between brain regions, traditionally examining connectivity between a seed region of interest (ROI) and the voxels of the whole rest brain. However, this study conducted a generalized PPI analysis, which is on ROI-to-ROI basis (Di et al., 2021), to yield a gPPI matrix across the whole brain instead of just a single seed region.

      Given the high frequency of Go trials in SST, it is common to treat Go trials as an implicit baseline in previous IMAGEN studies (D'Alberto et al., 2018; Whelan et al., 2012). Hence, we built a separate GLM for Successful stop trials, which included two task regressors (Failed and Successful stop trials) and 36 nuisance regressors.

      (6) Why did the author use PPI to construct the network, rather than the other similar methods, for example, beta series correlation (BSC)?

      Thanks for your question. PPI is an approach used to calculate the functional connectivity (FC) under a specific task (i.e., task-related FC). Although most brain connectomic research has utilized resting-state FC (e.g., beta series correlation), FC during task performance has demonstrated superiority in predicting individual behaviours and traits,  due to its potential to capture more behaviourally relevant information (Dhamala et al., 2022; Greene et al., 2018; Yoo et al., 2018). Specifically, Zhao et al. (2023) suggested that task-related FC outperforms both typical task-based and resting-state FC in predicting individual differences. Therefore, we chose to use task-related FC to predict sustained attention over time. We have updated it in the Introduction on Page 5.

      Introduction

      Although most brain connectomic research has utilized resting-state fMRI data, functional connectivity (FC) during task performance has demonstrated superiority in predicting individual behaviours and traits, due to its potential to capture more behaviourally relevant information (Dhamala et al., 2022; Greene et al., 2018; Yoo et al., 2018). Specifically, Zhao et al. (2023) suggested that task-related FC outperforms both typical task-based and resting-state FC in predicting individual differences. Hence, we applied task-related FC to predict sustained attention over time.

      (7) In the section of 'Correlation analysis between the network strength and substance use', the author just described that 'the correlations between xx and xx are shown in Fig5X', and repeated it three times for three correlation results. What exactly are the results? The author should describe the results in detail. And I am wondering whether there are scatter plots for these correlation analyses?

      We’d like to clarify the results in Fig. 5. Fig. 5 illustrates the significant correlations between behaviour and brain activity associated with sustained attention and Cigarette and cannabis use (Cig+CB) after FDR correction. Panel A shows the significant correlation between behaviour level of sustained attention and Cig+CB. Panels B and C show the correlations between brain activity associated with sustained attention and Cig+CB. While Panel B presents the brain activity derived from Go trials, Panel C presents brain activity derived from Successful stop trials. In response to your suggestion, we have described these results in detail on Page 9. We also have included scatter plots for the significant correlations, which are shown in Fig. 5 in Supplementary materials (Fig. S10).

      Results

      (6) Correlation between behaviour and brain to cannabis and cigarette use

      Figs. 5A-C summarizes the results showing the correlation between ICV/brain activity and Cig+CB per timepoint and across timepoints. Fig. 5A shows correlations between ICV and Cig+CB (Tables S14-15). ICV was correlated with Cig+CB at ages 19 (Rho = 0.13, P < 0.001) and 23 (Rho = 0.17, P < 0.001). ICV at ages 14 (Rho = 0.13, P = 0.007) and 19 (Rho = 0.13, P = 0.0003) were correlated with Cig+CB at age 23. Cig+CB at age 19 was correlated with ICV at age 23 (Rho = 0.13, P = 9.38E-05). Fig. 5B shows correlations between brain activity derived from Go trials and Cig+CB (Tables S18-19). Brain activities of positive and negative networks derived from Go trials were correlated with Cig+CB at age 23 (positive network: Rhop = 0.12, P < 0.001; negative network: Rhon = -0.11, P < 0.001). Brain activity of the negative network derived from Go trials at age 14 was correlated with Cig+CB at age 23 (Rhon = -0.16, P = 0.001). Cig+CB at age 19 was correlated with brain activity of the positive network derived from Go trials at age 23 (Rhop = 0.10, P = 0.002). Fig. 5C shows the correlations between brain activity derived from Successful stop and Cig+CB (Tables S18-19). Brain activities of positive and negative networks derived from Successful stop were correlated with Cig+CB at ages 19 (positive network: Rhop = 0.10, P = 0.001; negative network: Rhon = -0.08, P = 0.013) and 23 (positive network: Rhop = 0.13, P < 0.001; negative network: Rhon = -0.11, P = 0.001).

      (8) Lastly, the labels of (A), (B) ... in the figure captions are unclear. The authors should find a better way to place the labels in the caption and keep them consistent throughout all figures.

      Thank you for this valuable comment. We have revised the figure captions in the main text to ensure the labels (A), (B), etc., are placed more clearly and consistently across all figures.

      Reviewer #2 (Public Review):

      While the study largely achieves its aims, several points merit further clarification:

      (1) Regarding connectome-based predictive modeling, an assumption is that connections associated with sustained attention remain consistent across age groups. However, this assumption might be challenged by observed differences in the sustained attention network profile (i.e., connections and related connection strength) across age groups (Figures 2 G-I, Fig. 3 G_I). It's unclear how such differences might impact the prediction results.

      Thank you for your insightful comment. We’d like to clarify that we did not assume that connections associated with sustained attention remain completely consistent across age groups. Indeed, we expected that connections would change across age groups, due to the developmental changes in brain function and structure from adolescence to adulthood. Our focus was on the consistency of individual differences in sustained attention networks over time, recognising that the actual connections within those networks may change. However, we did show that there is some consistency in the specific connections associated with sustained attention over time. Notably, this consistency markedly increases when comparing ages 19 and 23, when developmental factors are less relevant. We support our reasoning above with the following analyses:

      (1) Supplementary materials (Pages 2 and 5), relevant sections highlighted here for emphasis.

      Method

      Comparison of predictive networks identified at one timepoint versus another

      Steiger’s Z value was employed to compare predictive performances of networks identified at different timepoints. This analysis involved comparing the R values derived from networks defined at distinct ages to predict ICV at the same age. For example, we compared the r values of brain networks defined at age 14 when predicting ICV at 19 (i.e., positive network: r = 0.25, negative network: r = 0.25, combined network: r = 0.28) with those R values of brain networks defined at age 19 itself (i.e., positive network: r = 0.16, negative network: r = 0.14, combined network: r = 0.16) derived from Go trials using Steiger's Z test (age 14 → age 19 vs. age 19 → 19). Similarly, comparisons were made between networks defined at age 14 predicting ICV at age 23 and those at age 23 predicting ICV at age 23 (age 14 → age 23 vs. age 23 → 23), as well as between networks defined at age 19 predicting ICV at age 23 and those at age 23 predicting ICV at age 23 (age 19 -> age 23 vs. age 23 -> age 23). These comparisons were performed separately for Go trials and Successful Stop trials.

      Results

      Comparison of predictive performance at different timepoints

      For positive, negative, and combined networks predicting ICV derived from Go trials at age 19, the R values were higher when using predictive networks defined at 19 than those defined at 14 (Z = 3.79, Z = 3.39, Z = 3.99, all P < 0.00071). Similarly, the R values for positive, negative, and combined networks predicting ICV derived from Go trials at age 23 were higher when using predictive networks defined at age 23 compared to those defined at ages 14 (Z = 6.00, Z = 5.96, Z = 6.67, all P < 3.47e-9) or 19 (Z = 2.80, Z = 2.36, Z = 2.57, all P < 0.005).

      At age 19, the R value for the positive network predicting ICV derived from Successful stop trials was higher when using predictive networks defined at 19 compared to those defined at 14 (Z = 1.54, P = 0.022), while the negative and combined networks did not show a significant difference (Z = 0.85, P = 0.398; Z = 2.29, P = 0.123). At age 23, R values for the positive and combined networks predicting ICV derived from Successful stop trials were higher when using predictive networks defined at 23 compared to those defined at 14 (Z = 3.00, Z = 2.48, all P < 3.47e-9) or 19 (Z = 2.52, Z = 1.99, all P < 0.005). However, the R value for the negative network at age 23 did not significantly differ when using predictive networks defined at 14 (Z = 1.80, P = 0.072) or 19 (Z = 1.48, P = 0.138).

      These results indicate that some specific pairwise connections associated with sustained attention at earlier ages, such as 14 and 19, are still relevant as individuals grow older. However, some connections are not optimal for good sustained attention at older ages. That is, the brain reorganizes its connection patterns to maintain optimal functionality for sustained attention as it matures.

      (2) Consistency of Individual Differences:

      We found individual differences in ICV were significantly correlated between the three timepoints (Fig. 1B). In addition, we calculated the correlations of network strength of predictive networks predicting sustained attention derived from Go trials and Successful trials between each timepoints. We found that the correlations of network strength for predictive networks (derived from Go trials and Successful trials) were also significant (all P < 0.003). We have updated these results in the main text (Pages 7-8) and Supplementary Materials (Table S7).

      (2) Cross-sectional brain connectivity

      In addition, we found that network strength of positive, negative, and combined networks derived from Go trials was significantly correlated between the three timepoints (Table S7, all P < 0.003).

      In addition, we found that network strength of positive, negative, and combined networks derived from Successful stop trials was significantly correlated between the three timepoints (Table S7, all P < 0.001).

      (3) Predictive networks across timepoints: Predictive networks defined at age 14 were successfully applied to predict ICV at ages 19 and 23. Similarly, predictive networks defined at age 19 were successfully applied to predict ICV at age 23 (Fig. 4). These results reflect the robustness of the brain network associated with sustained attention over time.

      (4) Dice coefficient analysis: We calculated the Dice coefficient to quantify the similarity of predictive networks across the three timepoints. Connections in the sustained attention networks were significantly similar from ages 14 to 23 (Table S13), despite relatively few overlapping edges over time (as discussed in Supplementary Materials on Page 6).

      (5) Global brain activation: Based on these findings, we indicate that sustained attention relies on global brain activation (i.e., network strength) rather than specific regions or networks (see also (Zhao et al., 2021)).

      In summary, brain network connections undergo change and are not completely consistent across time. However, individual differences in sustained attention and its network are consistent across time, as we found that 1) the brain reorganizes its connection patterns to maintain optimal functionality for sustained attention as it matures. 2) ICV and network strength of sustained attention network were significantly correlated between each timepoint. 3) Sustained attention networks identified from previous timepoints could predict ICV in the subsequent timepoint. 4) Dice coefficient analysis indicated that the edges in the sustained attention networks were significantly similar from ages 14 to 23. 5) Sustained attention networks function as a global activation, rather than specific regions or networks.

      (2) Another assumption of the connectome-based predictive modeling is that the relationship between sustained attention network and substance use is linear and remains linear over development. Such linear evidence from either the literature or their data would be of help.

      Thanks for your valuable suggestion. We'd like to clarify that while CPM assumes a linear relationship between brain and behaviour (Shen et al., 2017), it does not assume that the relationship between the sustained attention network and substance use remains linear over development.

      Our approach in applying CPM to predict sustained attention across different timepoints was based on previous neuroimaging studies (Rosenberg et al., 2016; Rosenberg et al., 2020), which indicated linear associations between brain connectivity patterns and sustained attention using CPM analysis. These findings support the notion of a linear relationship between brain connectivity and sustained attention. In this study, we performed CPM analysis to identify predictive networks predicting sustained attention, not substance use and used the network strength of these predictive networks to represent sustained attention activity.

      To examine the relationship between substance use and sustained attention, as well as its associated brain activity, we conducted correlation analyses and utilized a latent change score model instead of CPM analysis. This decision was informed by cross-sectional studies (Broyd et al., 2016; Lisdahl and Price, 2012) that consistently reported linear associations between substance use and impairments in sustained attention. Additionally, longitudinal research by (Harakeh et al., 2012) indicated a linear relationship between poorer sustained attention and the initiation and escalation of substance use over time.

      Given these previous findings, we assumed a linear relationship between sustained attention and substance use. Our analyses included calculating correlations between substance use and sustained attention, as well as its associated brain activity at each timepoint and across timepoints (Fig. 5). Furthermore, we employed a three-wave bivariable latent change score model, a longitudinal approach, to assess the relationship between substance use and behavirour and brain activity associated with sustained attention (Figs. 6-7). We have added more information in the Introduction to make it more clear on Page 6.

      Introduction

      Additionally, previous cross-sectional and longitudinal studies (Broyd et al., 2016; Harakeh et al., 2012; Lisdahl and Price, 2012) have shown that there are linear relationships between substance use and sustained attention over time. We therefore employed correlation analyses and a latent change score model to estimate the relationship between substance use and both behaviours and brain activity associated with sustained attention.

      (3) Heterogeneity in results suggests individual variability that is not fully captured by group-level analyses. For instance, Figure 1A shows decreasing ICV (better-sustained attention) with age on the group level, while there are both increasing and decreasing patterns on the individual level via visual inspection. Figure 7 demonstrates another example in which the group with a high level of sustained attention has a lower risk of substance use at a later age compared to that in the group with a low level of sustained attention. However, there are individuals in the high sustained attention group who have substance use scores as high as those in the low sustained attention group. This is important to take into consideration and could be a potential future direction for research.

      Thanks for this valuable comment. We appreciate your observation regarding the individual variability that is not fully captured by group-level analyses to some degree. Fig. 1A shows the results from a linear mixed model, which explains group-level changes over time while accounting for the random effect within subjects. Similarly, Fig. 7 shows the group-level association between substance use and sustained attention. We agree that future research could indeed consider individual variability. For example, participants could be categorized based on their consistent trajectories of ICV or substance use (i.e., keep decreasing/increasing) over multiple timepoints. We agree that incorporating individual-level analyses in the future could provide valuable insights and are grateful for your suggestion, which will inform our future research directions.

      The above-mentioned points might partly explain the significant but low correlations between the observed and predicted ICV as shown in Figure 4. Addressing these limitations would help enhance the study's conclusions and guide future research efforts.

      We have updated the text in the Discussion on Page 13:

      Discussion

      However, there are still some individual variabilities not captured in this study, which could be attributed to the diversity in genetic, environmental, and developmental factors influencing sustained attention and substance use. Future research should aim to explore these variabilities in greater depth to gain better understanding of the relationship between sustained attention and substance use.

      Reviewer #3 (Public Review):

      Weaknesses: It's questionable whether the prediction approach (i.e., CPM), even when combined with longitudinal data, can establish causality. I recommend removing the term 'consequence' in the abstract and replacing it with 'predict'. Additionally, the paper could benefit from enhanced rigor through additional analyses, such as testing various thresholds and conducting lagged effect analyses with covariate regression.

      Thank you for your comment. We have replaced “consequence” by “predict” in the abstract.

      Abstract

      Previous studies were predominantly cross-sectional or under-powered and could not indicate if impairment in sustained attention was a predictor of substance-use or a marker of the inclination to engage in such behaviour.

      Reviewer #3 (Recommendations For The Authors):

      (1) The connectivity analysis predicts both baseline and longitudinal attention measures. However, given the high correlation in attention abilities across the three time-points, it's unclear whether the connectivity predicts shared variations of attention across three time points. It would be insightful to assess if predictions at the 2nd and 3rd-time points remained  significant after controlling for attention abilities at the initial time point.

      Thanks for your comments. We performed the CPM analysis to predict ICV at the 2nd and 3rd timepoint, controlling for ICV at age 14 as a covariate. We found that controlling for ICV at age 14, positive, negative, and combined networks derived from Successful stop trials defined at age 14 still predicted ICV at ages 19 and 23. In addition, positive, negative, and combined networks derived from Successful stop trials defined at age 19 predicted ICV at age 23. In addition, positive, negative, and combined networks derived from Go trials defined at age 19 still predicted ICV at age 23, after controlling for ICV at age 14. However, positive, negative, and combined networks derived from Go trials defined at age 14 had lower predictive performances in predicting ICV at ages 19 and 23, after controlling for ICV at age 14. Notably, controlling for ICV at the initial timepoint did not significantly impact the performances of predictive networks derived from Successful stop trials. Accordingly, we have added this analysis and the results in the Supplementary Materials (Pages 3 and 5).

      Method

      Prediction across timepoints controlling for ICV at age 14

      To examine whether connectivity predictors shared variations of sustained attention across timepoints, we applied predictive models developed at ages 14 and 19 to predict ICV at subsequent timepoints controlling for ICV at age 14. Specifically, we used predictive models (including parameters and selected edges) developed at age 14 to predict ICV at ages 19 and 23 separately. First, we calculated the network strength using the gPPI matrix at ages 19 and 23 based on the selected edges identified from CPM analysis at age 14. We then estimated the predicted ICV at ages 19 and 23 by applying the linear model parameters (slope and intercept) obtained from CPM analysis at age 14 to the network strength. Finally, we evaluated the predictive performance by calculating the partial correlation between the predicted and observed values at ages 19 and 23, controlling for ICV at age 14. Similarly, we applied models developed at age 19 to predict ICV at age 23, also controlling for ICV at age 14. To assess the significance of the predictive performance, we used a permutation test, shuffling the predicted ICV values and calculating partial correlation to general a random distribution over 1,000 iterations.

      Results

      Predictions across timepoints controlling for ICV at age 14

      Positive and combined networks derived from Go trials defined at age 14 predicted ICV at ages 19 (r = 0.10, P = 0.028; r = 0.08, P = 0.047) but negative network did not (r = 0.06, P = 0.119). Positive network derived from Go trials defined at age 14 predicted ICV at age 23 (r = 0.11, P = 0.013) but negative and combined networks did not (r = 0.04, P = 0.187; r = 0.08, P = 0.056).  Positive, negative, and combined networks derived from Go trials defined at age 19 predicted ICV at age 23 (r = 0.22, r = 0.19, and r = 0.22, respectively, all P < 0.001).

      Positive, negative, and combined networks derived from Successful stop trials defined at age 14 predicted ICV at age 19 (r = 0.08, P = 0.036; r = 0.10, P = 0.012; r = 0.11, P = 0.009) and 23 (r = 0.11, P = 0.005; r = 0.13, P = 0.005; r = 0.13, P = 0.017) respectively. Positive, negative, and combined networks derived from Successful stop trials defined at age 19 predicted ICV at age 23 (r = 0.18, r = 0.18, and r = 0.17, respectively, all P < 0.001).

      (2) In the Results section, a significance threshold of p = 0.01 was used for the CPM analysis. It would be beneficial to test the stability of these findings using alternative thresholds such as p = 0.05 or p = 0.005.

      We appreciate this insightful comment. We appreciate the suggestion to test the stability of our findings using alternative significance thresholds. Indeed, we have already conducted CPM analyses using a range of thresholds, including 0.1, 0.05, 0.01, 0.005, 0.001, 0.0005, and 0.0001 (see Table S8 in supplementary Materials). The results were similar across different thresholds. Following prior studies (Feng et al., 2024; Ren et al., 2021; Yoo et al., 2018) which used P < 0.01 for feature selection, we chose to focus on the threshold of P < 0.01 for our main analysis. Following your suggestion, we have highlighted this in the Method section on Pages 17-18.

      Method

      2.6.1 ICV prediction

      The r value with an associated P value for each edge was obtained, and a threshold P = 0.01 (Feng et al., 2024; Ren et al., 2021; Yoo et al., 2018) was set to select edges.

      2.6.2 Three cross-validation schemes

      In addition, we conducted the CPM analysis using a range of thresholds for feature selection and observed similar results across different thresholds (See Supplementary Materials Table S8).

      (3) Could you clarify if you used one sub-sample to extract connectivity related to sustained attention and then used another sub-sample to predict substance use with attention-related connectivity?

      Thank you very much for the question. We used the same sample to extract the brain network strength and estimated the correlation with substance use using both the Spearman correlation and latent change score model across three timepoints. We controlled for covariates including sex, age, and scan site at the same time. Accordingly, we have clarified this in the Method section on Page 20. We note that the CPM analyses were conducted using cross-validation, plus a leave-site-out analysis.

      Method

      2.7.3 Correlation between network strength and substance use

      It is worth noting that all the correlations between substance use and sustained attention were conducted using the same sample across three timepoints.

      (4) Could you clarify whether you have regressed covariates in the lagged effects analysis of part 7?

      Thanks for this question. Yes, we confirmed that we controlled the covariates including age, sex and scan sites in the latent change score model. We have described them more clearly now in the Method section (Page 18).

      Method

      2.7.3 Correlation between network strength and substance use

      Additionally, cross-lagged dynamic coupling (i.e., bidirectionality) was employed to explore individual differences in the relationships between substance use and linear changes in ICV/brain activity, as well as the relationship between ICV/brain activity and linear change in substance use. The model accounted for covariates such as age, sex and scan sites.

      References:

      Broyd, S.J., van Hell, H.H., Beale, C., Yucel, M., Solowij, N., 2016. Acute and Chronic Effects of Cannabinoids on Human Cognition-A Systematic Review. Biol Psychiatry 79, 557-567.

      Chamberlain, S.R., Odlaug, B.L., Schreiber, L.R.N., Grant, J.E., 2012. Association between Tobacco Smoking and Cognitive Functioning in Young Adults. The American Journal on Addictions 21, S14-S19.

      Crean, R.D., Crane, N.A., Mason, B.J., 2011. An evidence based review of acute and long-term effects of cannabis use on executive cognitive functions. J Addict Med 5, 1-8.

      D'Alberto, N., Chaarani, B., Orr, C.A., Spechler, P.A., Albaugh, M.D., Allgaier, N., Wonnell, A., Banaschewski, T., Bokde, A.L.W., Bromberg, U., Buchel, C., Quinlan, E.B., Conrod, P.J., Desrivieres, S., Flor, H., Frohner, J.H., Frouin, V., Gowland, P., Heinz, A., Itterman, B., Martinot, J.L., Paillere Martinot, M.L., Artiges, E., Nees, F., Papadopoulos Orfanos, D., Poustka, L., Robbins, T.W., Smolka, M.N., Walter, H., Whelan, R., Schumann, G., Potter, A.S., Garavan, H., 2018. Individual differences in stop-related activity are inflated by the adaptive algorithm in the stop signal task. Hum Brain Mapp 39, 3263-3276.

      Dhamala, E., Yeo, B.T.T., Holmes, A.J., 2022. Methodological Considerations for Brain-Based Predictive Modelling in Psychiatry. Biological Psychiatry.

      Di, X., Zhang, Z.G., Biswal, B.B., 2021. Understanding psychophysiological interaction and its relations to beta series correlation. Brain Imaging and Behavior 15, 958-973.

      Dougherty, D.M., Mathias, C.W., Dawes, M.A., Furr, R.M., Charles, N.E., Liguori, A., Shannon, E.E., Acheson, A., 2013. Impulsivity, attention, memory, and decision-making among adolescent marijuana users. Psychopharmacology (Berl) 226, 307-319.

      Esterman, M., Rothlein, D., 2019. Models of sustained attention. Curr Opin Psychol 29, 174-180.

      Feng, Q., Ren, Z., Wei, D., Liu, C., Wang, X., Li, X., Tie, B., Tang, S., Qiu, J., 2024. Connectome-based predictive modeling of Internet addiction symptomatology. Soc Cogn Affect Neurosci 19.

      Greene, A.S., Gao, S., Scheinost, D., Constable, R.T., 2018. Task-induced brain state manipulation improves prediction of individual traits. Nature Communications 9, 2807.

      Harakeh, Z., de Sonneville, L., van den Eijnden, R.J., Huizink, A.C., Reijneveld, S.A., Ormel, J., Verhulst, F.C., Monshouwer, K., Vollebergh, W.A., 2012. The association between neurocognitive functioning and smoking in adolescence: the TRAILS study. Neuropsychology 26, 541-550.

      Hart, C.L., van Gorp, W., Haney, M., Foltin, R.W., Fischman, M.W., 2001. =. Neuropsychopharmacology 25, 757-765.

      Lawrence, N.S., Ross, T.J., Stein, E.A., 2002. Cognitive mechanisms of nicotine on visual attention. Neuron 36, 539-548.

      Lisdahl, K.M., Price, J.S., 2012. Increased marijuana use and gender predict poorer cognitive functioning in adolescents and emerging adults. J Int Neuropsychol Soc 18, 678-688.

      O'Halloran, L., Cao, Z.P., Ruddy, K., Jollans, L., Albaugh, M.D., Aleni, A., Potter, A.S., Vahey, N., Banaschewski, T., Hohmann, S., Bokde, A.L.W., Bromberg, U., Buchel, C., Quinlan, E.B., Desrivieres, S., Flor, H., Frouin, V., Gowland, P., Heinz, A., Ittermann, B., Nees, F., Orfanos, D.P., Paus, T., Smolka, M.N., Walter, H., Schumann, G., Garavan, H., Kelly, C., Whelan, R., 2018. Neural circuitry underlying sustained attention in healthy adolescents and in ADHD symptomatology. Neuroimage 169, 395-406.

      Potter, A.S., Newhouse, P.A., 2008. Acute nicotine improves cognitive deficits in young adults with attention-deficit/hyperactivity disorder. Pharmacol Biochem Behav 88, 407-417.

      Ren, Z., Daker, R.J., Shi, L., Sun, J., Beaty, R.E., Wu, X., Chen, Q., Yang, W., Lyons, I.M., Green, A.E., Qiu, J., 2021. Connectome-Based Predictive Modeling of Creativity Anxiety. Neuroimage 225, 117469.

      Rosenberg, M.D., Finn, E.S., Scheinost, D., Papademetris, X., Shen, X., Constable, R.T., Chun, M.M., 2016. A neuromarker of sustained attention from whole-brain functional connectivity. Nat Neurosci 19, 165-171.

      Rosenberg, M.D., Scheinost, D., Greene, A.S., Avery, E.W., Kwon, Y.H., Finn, E.S., Ramani, R., Qiu, M., Constable, R.T., Chun, M.M., 2020. Functional connectivity predicts changes in attention observed across minutes, days, and months. Proc Natl Acad Sci U S A 117, 3797-3807.

      Shen, X., Finn, E.S., Scheinost, D., Rosenberg, M.D., Chun, M.M., Papademetris, X., Constable, R.T., 2017. Using connectome-based predictive modeling to predict individual behavior from brain connectivity. Nat Protoc 12, 506-518.

      Valentine, G., Sofuoglu, M., 2018. Cognitive Effects of Nicotine: Recent Progress. Curr Neuropharmacol 16, 403-414.

      Verbruggen, F., Aron, A.R., Band, G.P.H., Beste, C., Bissett, P.G., Brockett, A.T., Brown, J.W., Chamberlain, S.R., Chambers, C.D., Colonius, H., Colzato, L.S., Corneil, B.D., Coxon, J.P., Dupuis, A., Eagle, D.M., Garavan, H., Greenhouse, I., Heathcote, A., Huster, R.J., Jahfari, S., Kenemans, J.L., Leunissen, I., Li, C.S.R., Logan, G.D., Matzke, D., Morein-Zamir, S., Murthy, A., Pare, M., Poldrack, R.A., Ridderinkhof, K.R., Robbins, T.W., Roesch, M.R., Rubia, K., Schachar, R.J., Schall, J.D., Stock, A.K., Swann, N.C., Thakkar, K.N., van der Molen, M.W., Vermeylen, L., Vink, M., Wessel, J.R., Whelan, R., Zandbelt, B.B., Boehler, C.N., 2019. A consensus guide to capturing the ability to inhibit actions and impulsive behaviors in the stop-signal task. Elife 8.

      Whelan, R., Conrod, P.J., Poline, J.B., Lourdusamy, A., Banaschewski, T., Barker, G.J., Bellgrove, M.A., Buchel, C., Byrne, M., Cummins, T.D., Fauth-Buhler, M., Flor, H., Gallinat, J., Heinz, A., Ittermann, B., Mann, K., Martinot, J.L., Lalor, E.C., Lathrop, M., Loth, E., Nees, F., Paus, T., Rietschel, M., Smolka, M.N., Spanagel, R., Stephens, D.N., Struve, M., Thyreau, B., Vollstaedt-Klein, S., Robbins, T.W., Schumann, G., Garavan, H., Consortium, I., 2012. Adolescent impulsivity phenotypes characterized by distinct brain networks. Nat Neurosci 15, 920-925.

      Yoo, K., Rosenberg, M.D., Hsu, W.T., Zhang, S., Li, C.R., Scheinost, D., Constable, R.T., Chun, M.M., 2018. Connectome-based predictive modeling of attention: Comparing different functional connectivity features and prediction methods across datasets. Neuroimage 167, 11-22.

      Young, J.W., Finlayson, K., Spratt, C., Marston, H.M., Crawford, N., Kelly, J.S., Sharkey, J., 2004. Nicotine improves sustained attention in mice: evidence for involvement of the alpha7 nicotinic acetylcholine receptor. Neuropsychopharmacology 29, 891-900.

      Zhao, W., Makowski, C., Hagler, D.J., Garavan, H.P., Thompson, W.K., Greene, D.J., Jernigan, T.L., Dale, A.M., 2023. Task fMRI paradigms may capture more behaviorally relevant information than resting-state functional connectivity. Neuroimage, 119946.

      Zhao, W., Palmer, C.E., Thompson, W.K., Chaarani, B., Garavan, H.P., Casey, B.J., Jernigan, T.L., Dale, A.M., Fan, C.C., 2021. Individual Differences in Cognitive Performance Are Better Predicted by Global Rather Than Localized BOLD Activity Patterns Across the Cortex. Cereb Cortex 31, 1478-1488.