10,000 Matching Annotations
  1. Dec 2025
    1. Still others have criticized journalists and reporters for a tendency toward reductive presentations of complex issues involving minorities, such as the religious and racial tensions fueled by the September 11 attacks. By reducing these conflicts to “opposing frames”—that is, by oversimplifying them as two-sided struggles so that they can be quickly and easily understood—the news media helped create a greater sense of separation between Islamic Americans and the dominant culture after September 11

      This is a great example.

    1. Many argue that because celebrities are “public figures,” the same privacy rights that protect the general public don’t apply.

      My opinion is that celebrities don't owe people the details of their personal lives. While there is an invitation for parasocial behavior due to the nature of being a public figure, they still have a right to privacy.

    1. it seeks to make money from whatever cultural elements it can throughout the world.

      If the definition of imperialism is the way one country asserts power over other countries, are endeavors to influence culture or profit from culture attempts to gain power?

    1. In short, while AI tools enhance surface-level writing mechanics, they cannot replace the analytical depth and originality developed through traditional instruction and personalized feedback from educators

      Although AI is helpful, it cannot replace what traditional instruction does

    2. The ethical concerns surrounding AI-assisted writing primarily revolve around plagiarism, originality, and academic integrity.

      Biggest concerns involving AI.

    3. Writing is more than producing grammatically correct, coherently structured sentences; it requires developing cognitive abilities, such as critical thinking, argument construction, and independent reasoning. Traditional instruction encourages students to practice these skills over time, reinforcing long-term retention by incorporating exercises, discussions, and feedback loops [21]. AI provides students with a polished product, but it does not require them to engage fully in the writing and learning process. Although AI enhances short-term writing proficiency, it does not foster long-term retention of writing skills to the same extent as traditional instruction [32].

      AI cannot do the most important parts of writing

    4. AI can also help with brainstorming to assist students in generating ideas and logically structuring their essays.

      AI can be helpful in the early stages of writing

    5. AI writing tools provide polished, well-structured responses but often lack genuine insight, critical analysis, and the writer’s unique voice.

      AI's lack of creativity

    6. While AI can provide well-structured content, it does little to encourage the cognitive effort necessary for students to form their own ideas and arguments. This lack of engagement can easily lead to a decline in students’ critical thinking skills [2,15,28].

      good quote or paraphrase

    7. Students no longer have to construct their arguments or structure their essays to explain their points of view; instead, they simply let the AI writing tool perform the work. This can ultimately diminish their creativity and critical thinking abilities

      Overreliance on AI can lead to a loss of creativity and critical thinking

    8. The applications “learn” from massive amounts of text data found on the internet and through other sources. They can understand context, syntax, and tone, enabling them to create coherent sentences and paragraphs in various styles [4]. Moreover, as these tools have advanced, they are increasingly becoming essential for content creation in multiple fields because they offer efficiency and scalability for businesses, educators and students, and other creative writing professionals

      More about what AI does and why people use it

    1. On the most basic level, much of media is language and culture based and, as such, does not necessarily translate well to foreign countries. Thus, media globalization often occurs on a more structural level, following broader “ways of organizing and creating media

      In her TedTalk "The Danger of the Single Story", Chimamanda Adichie shares how during her childhood, the language and content of the books, which were in English and centered American and British characters, she read had a massive influence on her. How will globalization impact people as we become more intertwined through our economies and the internet?

    1. “Every person who shall monopolize, or attempt to monopolize, or combine or conspire with any other person or persons, to monopolize any part of the trade or commerce among the several States … shall be deemed guilty of a felony

      It is interesting to learn about how these laws came to exist.

    1. Entire East Wing of White House to be demolished for Trump's $200M ballroom

      What is included is just the facts that the East Wing was going to be demolished to make way for the ballroom

  2. social-media-ethics-automation.github.io social-media-ethics-automation.github.io
    1. Ted Chiang. Will A.I. Become the New McKinsey? The New Yorker, May 2023. URL: https://www.newyorker.com/science/annals-of-artificial-intelligence/will-ai-become-the-new-mckinsey (visited on 2023-12-10).

      AI like nearly everything else will be used to maximize corporate profits, likely at the expense of the workers. Similar to consulting firms, it could make harmful decisions under the excuse of optimization.

    2. Britney Nguyen. Ex-Twitter engineer says he quit years ago after refusing to help sell identifiable user data, worries Elon Musk will 'do far worse things with data'. November 2022. URL: https://www.businessinsider.com/former-twitter-engineer-worried-how-elon-musk-treat-user-data-2022-11 (visited on 2023-12-10).

      The article goes over how and ex software engineer at Twitter was asked by a telco company to allow them to peer in on user data. After some back and fourth, the telco company made it clear that they wanted data to tell them everything about a particular user, and Twitter confirmed that something like that wouldn't break any of their guidelines. The engineer left shortly after, but he explains his worries about Elon Musk and his ownership and how it will likely be worse.

    3. Steve Krenzel [@stevekrenzel]. With Twitter's change in ownership last week, I'm probably in the clear to talk about the most unethical thing I was asked to build while working at Twitter. 🧵. November 2022. URL: https://twitter.com/stevekrenzel/status/1589700721121058817 (visited on 2023-12-10).

      This is a tweet by Steve Krenzel talking about how since the owners of twitter have switched, he can speak about the most unethical thing he has done for twitter. This not only reveals that twitter does unethical things but also that other social media sites most likely do the same, possibly violating our privacy and gathering very private data we don't want them to consume.

    4. Alfred Nobel. December 2023. Page Version ID: 1189282550. URL: https://en.wikipedia.org/w/index.php?title=Alfred_Nobel&oldid=1189282550 (visited on 2023-12-10).

      This leads to a page about Alfred Nobel, inventor of dynamite and the first person to recieve a nobel peace prize.

    5. Britney Nguyen. Ex-Twitter engineer says he quit years ago after refusing to help sell identifiable user data, worries Elon Musk will 'do far worse things with data'. November 2022. URL: https://www.businessinsider.com/former-twitter-engineer-worried-how-elon-musk-treat-user-data-2022-11 (visited on 2023-12-10).

      This article was eye opening to me. It highlights the moral challenge engineers face over being asked to violate user data. I think it is admirable that Steve left over this but it makes me think about all of them that didnt.

    6. Britney Nguyen. Ex-Twitter engineer says he quit years ago after refusing to help sell identifiable user data, worries Elon Musk will 'do far worse things with data'. November 2022. URL

      This article was very disturbing to me, but honestly not surprising. Steve Krenzel is a former twitter engineer and in this article it explains how he came out and said that he quit after an absurd request to violate users' data. A company wanted to know all sorts of information about users of twitter so that they could use it against competitors. Steve was trying to do this request in a way that still protected users' privacy, but the company kept getting angry and saying that they needed more. They even said they get much more information from other tech companies. Steve says the he is worried the Elon Musk will grossly mishandle users' data and I would agree. I think that all social media companies handle users' data poorly but him specifically will be a terrible person to have that power.

    7. Ted Chiang. Will A.I. Become the New McKinsey? The New Yorker, May 2023. URL: https://www.newyorker.com/science/annals-of-artificial-intelligence/will-ai-become-the-new-mckinsey (visited on 2023-12-10).

      This source discusses that AI functions less like a regular tool and rather like a global consulting force similar to that of McKinsey. Stating that AI is currently used to strengthen corporate power and increase inequality. Warning that right now, without changes, AI will benefit the corporate world more than an individual.

    8. [u2] Luddite. December 2023. Page Version ID: 1189255462. URL: https://en.wikipedia.org/w/index.php?title=Luddite&oldid=1189255462 (visited on 2023-12-10).

      Luddite is a word that opposes the industrialism and other new technology. I personally is not a Luddite, because the development of technology can help human's life, and the transportation will also develop more with it. The technology helped people more than it harm, so we have to support technology.

    9. Ted Chiang. Will A.I. Become the New McKinsey? The New Yorker, May 2023. URL: https://www.newyorker.com/science/annals-of-artificial-intelligence/will-ai-become-the-new-mckinsey (visited on 2023-12-10).

      I think Chiang's warning is necessary - many people have utopian expectations of AI, hoping it will liberate the workforce and make life easier. But the reality is likely to be - if no constraints are imposed and the system remains unchanged, AI will eventually become a tool for capital to enhance control, exploit workers, and exacerbate inequality.

    10. Eli Whitney. December 2023. Page Version ID: 1189351897. URL:

      This passage introduces Eli Whitney, an important American inventor, whose invention of the cotton gin greatly changed the early economic structure of the United States. It also pointed out the complex impact of this invention; although it enhanced production efficiency, it also strengthened and prolonged the slavery system in the United States.

    1. Thamus replied: […] this discovery of yours will create forgetfulness in the learners’ souls, because they will not use their memories; they will trust to the external written characters and not remember of themselves. The specific which you have discovered is an aid not to memory, but to reminiscence, and you give your disciples not truth, but only the semblance of truth; they will be hearers of many things and will have learned nothing; they will appear to be omniscient and will generally know nothing; they will be tiresome company, having the show of wisdom without the reality.

      This reminds me of how some people thought the invention of calculators would make people dumber and unable to do calculations themselves. However calculators just enabled people to learn harder math. I wonder what other inventions we judge too pessimistically at first.

    2. Fig. 21.3 Tweet parodying how tech innovator often do blatantly unethical things [u7]# Many

      I remember when this tweet was popular and how it was very much a commentary on not just unethical inventors, but specifically about Mark Zuckerberg and Meta trying to make the Metaverse, which was originally conceptualized by sci-fi author Neal Stephenson's novel Snow Crash. If I remember correctly the book showed the idea of the metaverse in a negative light and explored the consequences of having a virtual world, which the tweet parodies.

    3. Many people like to believe (or at least convince others) that they are doing something to make the world a better place,

      This is so true. Many people set out to do good only to faced with the corruption of the world. Money is often the cause of this corruption and when money loses value it becomes power and that is how we end up with people who set out to do good and end up doing the worst.

    1. In my opinion, Ms. Tebbit's role felt much more significant than the article suggests. I wonder how involved she truly was in the execution of Timothy's actions, as she appeared to be almost all-knowing. The film seems to suggest that she knows more about the love potion and the magic that occurs more than she lets on. It would be interesting to see more about if/why she set the events in motion that caused the town to question their prejudices, because the article frames her more as a quirky teacher than possibly the cause of everything that occurred.

    2. As the article describes how "the persons involved slowly began to recognize their homophobia," it makes me wonder what the result would be if this were to happen in real life. Obviously the end result of the film was incredibly positive, and I wonder how much of that would carry over to real life biases.

    3. I love how the movie incorporated Shakespeare's original text by warping it into songs. As the article says, the language served as a way to echo the original themes and stay true to the writing, but also managed to convey a completely different meaning. I'm glad that we were able to see where the lyrics came from in class before watching the movie, because it allowed an entirely different watching experience.

    4. I wonder how the class would view the ethical dilemma on this. Is it okay to forcibly cause someone to see life thorough your point of view in a way if it is for the greater good? And could someone also argue that although obviously it was a happy ending and good that homophobia disappeared in his town, is it technically for the greater good if Timothy initially had self serving purposes (causing Jonathan to like him, getting back at his bullies)?

    5. I think it's worthy to note that even though she didn't accept it at the beginning, it seemed to be becuase of judgement from others and preconceived beleifs, but even with such, she still supported him and made his wings

    6. The historical context surrounding Elizabethan London strengthens the argument by showing how deeply rooted homophobia is. It's very interesting to see and something that I didn't know, about how queerness was viewed in Shakespeare's time. From what the article says, it sounds like his plays brought more awareness and normality to queer themes, which makes Were The World Mine even more of a commentary on social norms, since it bridges the past and present. This could be something very interesting to bring up to the class.

    7. I think the comparison between Timothy and Puck is very interesting. While both use the love potion in a way that causes chaos, Puck does so more on accident in the original Midsummer play. Timothy on the other hand uses the potion as a catalyst to both expose and correct the town's homophobia. Because his actions are more purposeful, this film becomes more of a discussion about society rather than simply a comedy. I wonder if we could have the class compare the two characters and their motivations, as well as the end results.

    1. As information and media move online, those without ready access to the Internet are increasingly being left behind.

      This issue was highlighted during the pandemic, when students were forced to rely on technology to go to school. Sone students and families didn't have access to technology while others didn't have reliable internet.

    1. This business has proven extremely productive; the bulk of Google’s revenue comes from advertising even as it gives away services such as email and document sharing

      How can we say that Google is "giving away these services" when they are selling out time and attention to the companies that advertise with google AdSense?

    1. Research papers generally follow the same basic structure: an introduction that presents the writer’s thesis, a body section that develops the thesis with supporting points and evidence, and a conclusion that revisits the thesis and provides additional insights or suggestions for further research.

      How a research paper is normally written

    2. This section explains that the way you write your research paper depends on its genre and purpose whether you're informing, persuading, or analyzing and that you need to keep that goal in mind while drafting.

    1. Universal is owned by NBC, which was in turn owned by GE and now Comcast; Sony Music is owned by the eponymous Japanese technology giant; Warner Music Group, although now its own entity, was previously under the umbrella of Time Warner; and the EMI Group is owned by a private investment firm.

      The legacy of GE is so sad and a prime example of the failure of conglomerates. Jack Welch drove General Electric into the ground, failing its workers and customers.

    2. The book publishing industry is basically an oligopoly; the top 10 trade publishers made up 72 percent of the total market in 2009, with the top five alone comprising 58 percent of this

      I didn't realize that this is how the publishing industry operates! I wonder what the stats are today.

    1. books and scholarly articles. Academic books generally fall into three categories: (1) textbooks written with students in mind, (2) monographs which give an extended report on a large research project, and (3) edited-volumes in which each chapter is authored by different people. Scholarly articles appear in academic journals, which are published multiple times a year in order to share the latest research findings with scholars in the field.

      Main forms of sources that are reliable

    2. The purpose of an informative essay, sometimes called an expository essay, is to educate others on a certain topic. Typically, these essays aim to answer the five Ws and H questions: who, what, where, when, why, and how.

      Informative essay should focus on the five W's and How.

    3. . The Informative Research Report is a report that relays the results of a central research question in an organized manner through more formal sources. These resources could include Google Scholar, library catalogs and academic article databases, websites of relevant agencies, and Google searches using (site: *.gov or site: *.org). A report is written from the perspective of someone who is seeking to find specific and in-depth information about a certain aspect of a topic.

      Informative report is meant to teach something from reliable sources.

    1. In the third quarter of 2008, the average American watched 142 hours of television per month, an increase of five hours per month from the same quarter the previous year. Internet use averaged 27 hours per month, an increase of an hour and a half between 2007 and 2008.

      According to globalstatistic.com, "people now spend more than twice as long using the internet daily as they do watching television. However, this comparison becomes more complex when considering that streaming TV shows and movies account for a significant portion of daily internet time, and over 30% of users access connected content through their TV sets, blurring the traditional boundaries between these media formats". It is interesting seeing how the internet and TV have coevolved, especially post pandemic.

    2. were allowed to encrypt their satellite feeds so that only people who purchased a decoder from a satellite provider could receive the channel.

      I'm curious what would have happened if the 1984 Cable Act hadn't been successful. Would "open skies" policies be more commonplace and the rights of consumers prioritized?

    1. Trump announced Monday that construction crews had begun tearing down part of the East Wing to make room for his privately funded White House ballroom, a project that's been in the works for months.

      Framing it as something not shocking or new

    2. the people's house.

      Second time this has been mentioned, in this context they are quoting the "liberal media". Saying Trump is taking away the people's house.

    3. 'Donald Trump is literally destroying the people's house,' Rachel Maddow says

      What is left out: Why is it being demolished, what all is being torn down?

    4. Democrats, liberal media figures fume over Trump demolishing White House East Wing for ballroom

      Headline: This tells the reader about the reaction to this news event. It uses the word fume to show that liberals are angry, which plays into the news values of conflict.

    1. In a corresponding move, the film chips awayat the integrity of Dream’s language by breakingoff some of its defining components and reassign-ing them to other genres and other media

      The author argues that the film weakens the original meaning and power of A Midsummer Night’s Dream by taking parts of Shakespeare’s language and repurposing them in musical numbers, fantasy sequences, or other cinematic forms

    2. As with the Neil story-line in Dead Poets Soci-ety, the focal point of the final scenes is Puck’sepilogue and the audience response. The familiarsequence of shots draws all the film’s charactersinto easy harmony. All of Timothy’s doubters anddetractors, together with his supporters, watch inapproving wonder. Donna smiles from the audi-ence and fields compliments about her son. Thecast engages in mutual congratulations, realizingthat they have delivered a first-rate performance,and the audience’s enthusiastic approval grows to astanding ovation when Timothy, like Neil, appearsfor a final bow.

      Were the World Mine uses the same cinematic pattern seen in Dead Poets Society: a climactic performance followed by an overwhelmingly positive audience reaction. The final scenes create a sense of harmony and acceptance, showing everyone supporters and former doubters united in admiration for Timothy. The applause and standing ovation serve as a symbolic confirmation that he has been validated and embraced by his community, mirroring Neil’s emotional moment on stage in Dead Poets Society.

    3. That is, while nobody remains under the effects ofPuck’s potions, nobody is returned to the preciseposition he or she formerly occupied. Homosex-uality is not the new standard, but it is accommo-dated in the new understanding of masculinity.

      After the chaos ends, the characters don’t go back to their original, rigidly homophobic attitudes. While the town doesn’t become entirely queer, it becomes more accepting, and masculinity is redefined to allow space for homosexuality

    4. he filmreturns to the genre conventions sketched here inmy brief accounts of Shakespeare in Love and DeadPoet’s Society. Despite its various departures fromthe formula, it is finally a traditional, which is tosay ‘straight’, saved-by-Shakespeare story

      Story still tells a conventional “saved-by-Shakespeare” story one that reinforces a traditional, heterosexual resolution rather than fully embracing a radical or transformative queer narrative.

    5. in films of this genre, and particularly in Were theWorld Mine, is not the rich, famously flexible con-tent that allows Shakespeare to be mobilized onbehalf of potentially contradictory positions, butthe archaic syntax and diction that give the linesa ‘Shakespeareness

      films use Shakespeare more for his recognizable “Shakespeareness” than for the real substance of his ideas

    6. And together, the films demonstrate the flexibil-ity of Shakespeare to accommodate the needs anddesires of whoever is speaking the language

      Shakespeare’s works are adaptable and can be reshaped to fit the intentions and perspectives of anyone who uses his language

    1. By bulldozing the structure, Trump is effectively erasing women from the history of the White House.

      News Values: conflict, framing that this is an attact on womens history

    2. No Shame The demolition of the White House’s East Wing is the perfect metaphor for what Trump is doing to the United States

      Headline: "No shame" is calling Trump out and criticizing the decision. Comparing the destruction to the White House to the (in their opinion) destruction of the country.

    1. Both incidents happened in the early hours of New Year's Day. Both men served in the US armed forces - including at the Fort Liberty (formerly Fort Bragg) military base in North Carolina - and both completed a tour in Afghanistan. Both men also rented the vehicles they used through a mobile car rental application called Turo.

      This does call into question wether the two a=incidents were connected with already having many links between the two.

    2. "I'm comfortable calling it a suicide with a bombing that occurred immediately after," Sheriff McMahill said.

      Suicide had not been mentioned in prior articles or at least the term had not been.

    3. drove it more than 800 miles

      In prior articles the long drive had not been questioned as much until now in this article which raises the valid question of why so far?

    4. Four unanswered questions after Las Vegas Cybertruck explosion

      This headline is largely different in contrast to the other articles and takes more of a review of event approach then simply reporting the event that occured.

    1. Public Broadcasting Service (PBS) developed out of a report by the Carnegie Commission on Educational Television, which examined the role of educational, noncommercial television on society. The report recommended that the government finance public television in order to provide diversity of programming during the network era—a service created “not to sell products” but to “enhance citizenship and public service.”

      I believe that PBS plays a very important role in providing educational content. I learned so much from PBS programs as a child.

    1. he said the country needed to “rally around” him and Tesla CEO Elon Musk.

      Different take now on the symbolism of exploding the tesla in front of the trump hotel.

    2. He left cellphone notes saying he needed to “cleanse” his mind “of the brothers I’ve lost and relieve myself of the burden of the lives I took.”

      This note was spoken about in earlier articles was not mentioned to be on a phone so that is new within the story as well the content of the note appeared different when given more of the notes content.

    3. Trump hotel in Las Vegas used generative AI including ChatGPT to help plan the attack, Las Vegas police said Tuesday.

      Using attribution to make sure this does not appear to be the readers ideation.

    4. Generative AI used to help plan Tesla Cybertruck explosion outside Trump hotel in Las Vegas, police say

      This article comes out 6 days after the initial explosion but shows a different perspective of the planning of the explosion in relation to an earlier news story posted again by PBS.

    1. He was awarded a total of five Bronze Stars, including one with a valor device for courage under fire, a combat infantry badge and an Army Commendation Medal with valor.

      Speaking to the suspects awards.

    2. Livelsberger apparently harbored no ill will toward President-elect Donald Trump, Clark County sheriff’s officials said.

      Mentioned in the CNN article that he even openly supported Trump.

    3. “cleanse my mind” of the lives lost of people he knew and “the burden of the lives I took.”

      This shows a different light to the story that the soldier who commited this crime had to do this for the lives he took.

    4. Soldier who died in Cybertruck explosion left note saying it was a ‘wakeup call’ for country’s ills

      This stories headline now looks to an entirely different reason for the attach where previous articles had not addressed the reason for the attack.

    1. Sources: the sources in the story shape it as a controversial act by centering on a historian who has an opinion of loss and frustration with it. She states that Melania Trump is absent, and the story does not include her voice. The story is constructed to emphasize what is being destroyed rather than emphasizing the ballroom being built.

    2. Katherine Jellison, a historian at Ohio University and a scholar of first ladies.

      Irie Senter is the author, but Katherine Jellison is interviewed for the story. A historian and scholar of first ladies. This matters because she gives an expert opinion on first ladies and the East Wing's connection and importance to first ladies.

    3. Why the East Wing Destruction Fallout May Linger

      Headline: The word destruction is a loaded word that signifies something being destroyed and cannot be replaced. "May Linger" suggests uncertainty in how long it will take for the fallout to be over. This word choice highlights that this is a very negative action and is a controversial subject.

    1. Mayorkas told CNN’s Wolf Blitzer. There’s no evidence that anyone else was involved in orchestrating the Cybertruck incident, Mayorkas said. PlayCaption Styling Settings Police release video showing Cybertruck explode outside Las Vegas Trump Hotel 02:19 • Source: CNN Police release video showing Cybertruck explode outside Las Vegas Trump Hotel 02:19 Livelsberger was a member of the 10th Special Forces Group, which has members

      Given video of the attack which brings again an emotional appeal to show just how devastating the attack was.

    2. driver’s motive remains unclear and authorities are not aware of any connection “to any terrorist organizations around the world.”

      This one uses explanation of sources well to establish this is not the writers stance on this issue.

    3. While McMahill and others officials only referred to the driver as the “subject” and “person of interest,” they showed a photo of Livelsberger and gave his name at the news conference.

      This is largely important fact within this developing news story at the time. Why would law enforcement show the image of Livelsberger but not identify him with his name?

    4. though did not specifically say he died in the Vegas explosion.

      This is something to mention and seems odd that the Army would not say he passed in the explosion.

    5. Livelsberger was an Army Special Forces operations master sergeant, a senior enlisted rank, according to four US officials. He was on active duty in Germany with the 10th Special Forces Group but was on leave at the time of the incident, three officials said.

      This brings in an emotional appeal to humanize the suspect in a way.

    6. We notice the strategic use of images here at the beginning of this video especially them being placed at the beginning of the article. The left image looks to be the explosive material recovered from the scene. Then the right photo appears to be the suspects passport photo.

    7. Driver who exploded Tesla Cybertruck at Trump hotel in Las Vegas was an active-duty Army Green Beret,

      With this headline we learn that the driver was a Army Green-Beret.

    1. Video played at the news conference showed a load of fireworks-style mortars, gasoline cans and camping fuel canisters in the back of the truck.

      Would have mentioned which exact news confrence.

    2. according to an official

      They are making sure to connect their information about the story back to their sources. This was discussed in class to be elements to a good news article because they are establishing credibility.

    3. looking into any possible connections to the deadly attack in New Orleans earlier Wednesday but had not yet discovered any.

      Based off the proximity of these two events because they occured very quickly after one another.

    4. Las Vegas police looking for links between Cybertruck blast and New Orleans attack

      As spoken about during class headlines can tell you a lot about the way a story will sway and with its short nature writers often have to use metaphors to get their point across in as little words as possible.

    1. Muslims were often engaged in long-distance commerce, and even when they were not, the requirement to make a pilgrimage to Mecca (the Hajj) once during their lifetime encouraged followers of Islam to be more interested in travel and curious about the world beyond their home regions.

      It’s interesting that the Hajj encouraged Muslims to travel and explore beyond their regions. I find it fascinating how religion and commerce together promoted curiosity about the wider world.

    2. Specifically, in 1099 when crusaders like Godfrey of Bouillon captured Jerusalem from the Fatimid Caliphate and established the Crusader States, they discovered sugar production as a profitable local industry in the city and surrounding regions.

      It’s interesting that when the Crusaders captured Jerusalem in 1099, they discovered sugar production as a profitable industry. I find it fascinating how economic opportunities influenced their actions.

    3. Ferdowsi completed the Persian epic Shahnameh, preserving pre-Islamic Iranian legends. In Cairo, the Al-Azhar Mosque became one of the world's oldest universities and a center for Islamic learning.

      It’s fascinating that Ferdowsi preserved pre-Islamic Iranian legends in the Shahnameh. I also find it interesting that Al-Azhar Mosque became a major center of Islamic learning.

    4. The Sunni Berber Almoravids unified North Africa and invaded al-Andalus in 1086, halting Christian advances at the Battle of Zallaqa and imposing stricter Islamic practices.

      It’s fascinating that the Almoravids unified North Africa and stopped Christian advances at Zallaqa. I find it interesting how they also enforced stricter Islamic practices.

    5. In the Islamic world, the centralized power of the Abassid Caliphate was beginning to give way to more regionally-based dynasties, at the same time scientific and philosophical production was peaking.

      It’s interesting that while the Abbasid Caliphate’s central power was weakening, science and philosophy were flourishing. I find it fascinating how intellectual life can thrive even during political change.

    6. Nearly the entire Umayyad family was then massacred at a "Banquet of Blood" in Palestine (except one prince who escaped to Cordoba), and Abu al-ʿAbbas was proclaimed caliph in Kufa mosque, beginning the Abbasid Empire.

      It’s shocking that nearly the entire Umayyad family was killed at the “Banquet of Blood,” with only one prince escaping. I find it fascinating how this event led to Abu al-ʿAbbas becoming caliph and the start of the Abbasid Empire.

    7. The Umayyads were overthrown in a carefully-orchestrated coup that drew on this resentment, as well as that of Shiites angered by their violent suppression of revolts like Karbala in 680, which I mentioned previously.

      It’s interesting that the Umayyads were overthrown through a coup fueled by widespread resentment and Shiite anger over events like Karbala. I find it fascinating how these tensions finally led to their downfall.

    8. But the Umayyads had their opponents, and many Muslims resented their rule. Non-Arab Muslims were taxed at a higher rate and barred from holding high office. The caliphate was treated as an inherited kingship rather than an elected office. And the lavish court in Damascus clashed with the ascetic ideals of early Islam.

      It’s interesting that many Muslims opposed the Umayyads because of unfair taxes, limited opportunities, and lavish courts. I find it fascinating how these issues clashed with early Islamic ideals.

    9. An example of Umayyad architecture is the Al-Aqsa mosque in Jerusalem, which is the world's oldest surviving work of Islamic architecture and the oldest mosque after the Al-Haram in Mecca

      It’s fascinating that the Al-Aqsa Mosque is the world’s oldest surviving Islamic architecture after Al-Haram. I find it interesting how it has stood for so many centuries.

    10. The Canal connected the Yellow and Yangtze Rivers, allowing goods to flow safely between the north and south without being endangered by coastal pirates.

      It’s fascinating that the Canal connected the Yellow and Yangtze Rivers. I find it interesting how it made trade safer by avoiding coastal pirates.

    11. Theodoric imagined himself as the restorer and protector of "Roman" order in the west; Justinian saw him as just another barbarian.

      It’s interesting that Theodoric saw himself as restoring Roman order while Justinian viewed him as a barbarian. I find it fascinating how perspective shaped their reputations. I find it also funny

    12. This forced other Germanic tribes such as the Vandals, Suebi, and Burgundians, to move westward, across the Rhine into Roman Gaul.

      It’s interesting how the movement of one group pushed other Germanic tribes like the Vandals, Suebi, and Burgundians into Roman Gaul. I find it fascinating how one migration could trigger a chain reaction.

    13. The Emperor, Honorius, had moved his government from Milan to Ravenna on the Adriatic coast in 402; so Rome was poorly defended.

      It’s surprising that Emperor Honorius moved the government to Ravenna, leaving Rome poorly defended. I find it interesting how this decision affected the city’s safety.

    14. Especially under Attila, they preferred quick raids to plunder wealth and extort tribute, rather than wasting their time capturing cities they lacked the administrative structures to govern.

      It’s interesting that Attila’s forces focused on quick raids instead of capturing cities. I find it fascinating how their strategy relied on speed and plunder rather than long-term control.

    15. Even farther east, the Chinese Empire had fallen into a chaotic period with the end of the Han Dynasty in 220 CE. In the middle decades of the third century, three kingdoms rose (Wei, Shu, and Wu), which were followed by unification for about fifty years under the Western Jin, with a capital in Chang'an

      It’s interesting that China entered a chaotic period after the Han Dynasty ended. I find it fascinating how three kingdoms rose before briefly unifying under the Western Jin.

    16. While the Council of Nicaea was a significant step forward for the legitimacy of the church in Roman society, in 325 christians were still only about 10% to 20% of all Romans. Constantine continued to favor the church, funding churches such as the Basilica of St. Peter in Rome, exempting clergy from taxes, and appointing Christians to high office. But it wasn't until 380 that Emperor Theodosius declared Nicene Christianity the official religion of the empire and banned pagan worship.

      It’s interesting that even after the Council of Nicaea, Christians were still only 10–20% of the population. I find it fascinating how Constantine supported the church, but Christianity didn’t become the official religion until Theodosius in 380.

    17. By 250 CE, between 1% and 5% of the empire's 60 million people were Christians (or, up to 3 million).

      It’s surprising to learn that up to 3 million people, or 1–5% of the empire, were Christians by 250 CE. I find it interesting how quickly the faith was growing.

    18. The Senate, fearing mob violence, granted the conspirators amnesty but also pardoned Caesar's allies, whom they had been planning to prosecute.

      It’s surprising that the Senate gave amnesty to both the conspirators and Caesar’s allies. I find it interesting how they tried to prevent mob violence while balancing justice.

    19. The princess, Liu Xijun, became famous in Chinese literature as the author of a poem called the "Song of Sorrow":

      It’s interesting that Princess Liu Xijun is remembered in Chinese literature for her poem, the "Song of Sorrow." I’m curious to read it and see the emotions she expressed.

    20. The envoy was captured by the Xiongnu and held for a decade, but he escaped and completed his mission, traveling over 12,000 miles through modern Xinjiang and Uzbekistan. He returned to China in 126 BCE with maps of thirty-six kingdoms, exotic goods, and credible stories of an interconnected world beyond the Pamir Mountains.

      It’s amazing that the envoy survived a decade of captivity and still completed his mission. I find it fascinating that he traveled over 12,000 miles and brought back maps, goods, and stories of faraway lands.

    21. . His conquests added vast territories, wealth, and loyal legions, and the Commentarii reveals Caesar’s strategic mind. Caesar’s conquests reshaped Roman politics but sparked civil war

      It’s fascinating to see how Caesar’s conquests brought wealth and power while showing his strategic mind. I’m especially interested in how they reshaped Roman politics and led to civil war.

    22. Carthage paid a heavy reparation (3,200 talents of silver) and the Roman navy seized the islands of Corsica and Sardinia soon after.

      It’s surprising to learn that Carthage had to pay 3,200 talents of silver and lose Corsica and Sardinia to Rome. This makes me interested in how harshly Rome enforced its power.

    23. This paradoxical defiance-as-worship surprises Rama and because he is actually Vishnu, when he kills the demon, Ravana achieves moksha.

      Learning how Ravana’s defiance is actually seen as a kind of worship really surprises me, and I find that idea fascinating. It’s even more interesting that because Rama is Vishnu, Ravana’s death leads him to moksha.

    24. Xenophon had spent the war fighting as a mercenary for Persia and had been exiled from Athens as a result; so his account is considered a bit less impartial.

      Learning that Xenophon fought as a Persian mercenary and was exiled from Athens helps me see why his account might be less impartial. It actually makes me more curious about how his experiences shaped the way he wrote history.

    25. n which the "Hellenes" imagined themselves to have been the true Greeks, reclaiming their ancestral homelands from barbarians.

      These stories help me understand how the Hellenes saw themselves as the real Greeks returning to take back their original lands. Learning this shows me how myth and identity were closely connected in ancient Greece.

    26. Myths of returning Trojan War heroes like Homer's Odyssey included the Nostoi (Returns), telling the stories of Agamemnon, Menelaus, Nestor, and Orestes. Other legends describe the adventures of the Trojan prince Aeneas and the Greek hero Ajax.

      These myths show me how often Greek stories use the idea of returning to reclaim something that was lost. Learning about Jason, the Seven Against Thebes, and the Heracleidae helps me see how important this theme was to ancient Greek culture.

    1. Muslim trade networks expanded across the Sahara, connecting North and sub-Saharan Africa, facilitating the spread of Islam to West African kingdoms such as Ghana. The kingdom known as Wagadou or the land of herds in the local Soninke language, was a prosperous empire that had become known to Islam in accounts by Arab geographers like al-Fazari and Ibn Hawqal in the ninth and tenth centuries. Ghana's wealth was based on trade in gold, salt, ivory, and slaves. Although its power was beginning to wane by the eleventh century, it would be absorbed around 1240 into the even more powerful Mali Empire, which we will cover in the next chapter.

      It’s fascinating how Muslim trade networks spread Islam and connected North and sub-Saharan Africa. I also find it interesting that Ghana became so wealthy through trade in gold, salt, ivory, and slaves before eventually being absorbed into the Mali Empire.

    1. It is our responsibility to shine our OT light and share with others about the impact our profession has

      Suggested grammatical revision:

      It is our responsibility to shine our occupational therapy light and share about the impact of our profession with others.

      Take it or leave it!

    1. . Ali ruled for another five years, moving his capital to Mesopotamia, until he too was assassinated in 661; leading to the establishment of the Umayyad Caliphate that established its base in Damascus and solidified Sunni dominance.

      It’s interesting that Ali ruled for five years before his assassination. I find it fascinating how his death led to the Umayyad Caliphate and the rise of Sunni dominance.

  3. ot27grad.weebly.com ot27grad.weebly.com
    1. But beginning in the 1st century BCE, the kingdom of Aksum (or Axum) emerged during the decline of Ptolemaic Egypt and flourished as a trade center that linked Rome, India, Arabia, and sub-Saharan Africa.

      It’s fascinating that Aksum rose as Ptolemaic Egypt declined and became a major trade hub. I’m really interested in how it connected so many regions across Africa and beyond.

    1. At its height, the empire had a population of about 15 million and produced wealth similar to that of Rome at the same time.

      It’s amazing that the empire had around 15 million people and wealth comparable to Rome. I find it interesting to imagine how prosperous it must have been.

    1. This list has received positive feedback. It seems to have struck a nerve. Sending it to coincide with Monday morning to help folks navigate their week.Take care of yourself. Your brain is working overtime—all the time. Practice “radical” recovery.You may spend a lot longer thinking about things than most people. Pace your delivery.If you go deep first, and then simplify…keep in mind that you don’t need to show all of your work.Your default description of (almost) any problem will be too threatening/overwhelming.Do your deepest thinking with co-conspirators (not the people you’re trying to influence).Informal influence is often not formally recognized. Prepare mentally for this.The people you’re trying to influence spend 98% of their day overwhelmed by business as usual.Remember to also do the job you were hired to do (if you don’t you’ll be easier to discount).Seek “quick wins”, but know that most meaningful things will take a while.Some things take ages to materialize. It is discontinuous, not continuous.Make sure to celebrate your wins. They will be few and far between, so savor the moment.The people who support you in private may not be able to support you in public. Accept that.Hack existing power structures—it’s much easier than trying to change them.Consider becoming a formal leader. It’s harder in many ways, but you’ll have more leverage. What’s stopping you?In lieu of being a formal leader, make sure to partner with people who actually “own” the area of change.Watch out for imposing your worldview on people. Have you asked about what people care about?.You’ll need a support network. And not just a venting network. Real support.“Know when to fold ‘em”. Listen to Kenny Rogers The Gambler. Leave on your own terms.Don’t confuse being able to sense/see system dynamics, with being about to “control” them. You can’t.Grapple with your demons, and make sure not to wrap up too much of your identity in change.
    1. White House press secretary Karoline Leavitt responded to multiple questions about the lack of oversight for the project Thursday, saying Trump is following legal precedent. She also held up several images of past renovation projects that she said demonstrated how the president’s decision on the East Wing aligned with actions other presidents have taken.

      Sources: Karoline Leavitt, a bias source as her job is to defend the presidents decisions. She frames the decsion by emphasizing legal precedent and "held up serveral images" to use visual evidence that there were other revovations to normalize what is occuring

    2. But the images that emerged this week of construction equipment demolishing the facade have prompted outrage over the lack of oversight of the project and raised questions about its scope.

      This line uses key words to construct the story, stating "prompted outrage" indicates a strong emotional reaction to the news event. "Lack of oversight" suggests that this is occurring with to regard for the interest of the people or preservation.

    3. The East Wing of the White House, the traditional headquarters for the first lady and her staff as well as the public entrance to the “people’s house,” was largely torn down this week to make way for President Donald Trump’s new $300 million ballroom.

      The news values invoked here are prominence, due to the President being involved, conflict as many are opposed to this and unusalness as it isn't everyday that the White House is being torn down.

    4. The East Wing of the White House, the traditional headquarters for the first lady and her staff as well as the public entrance to the “people’s house,” was largely torn down this week to make way for President Donald Trump’s new $300 million ballroom.

      This frames the story in a way by including historical significance by using the word "traditional". The phrase "peoples house" reminds readrs that what is being torn down belongs to the nation, not just the President.

    1. With the outbreak of World War II, many companies, including RCA and General Electric, turned their attention to military production. Instead of commercial television sets, they began to churn out military electronic equipment. In addition, the war halted nearly all television broadcasting; many television stations reduced their schedules to around 4 hours per week or went off the air altogether.

      Many new technological advancements are a direct product of or inspired by products made for the military, for example food and clothes.

    2. Formerly known as Community Antenna Television, or CATV, cable television was originally developed in the 1940s in remote or mountainous areas, including in Arkansas, Oregon, and Pennsylvania, to enhance poor reception of regular television signals. Cable antennas were erected on mountains or other high points, and homes connected to the towers would receive broadcast signals.

      Access to technology can expand or limit a person's ability to engage with media and culture.

    1. coverage, which leads to the debate over media bias

      Media coverage can influence public knowledge and opinion about a subject, event, or individual. Some stories go unnoticed because there is little coverage while others become over saturated.

    1. A 56% majority of Americans oppose the construction of the ballroom, including 45% who "strongly" oppose it, according to the poll taken in October.

      This is framing, the approach of this is to highlight the majority of Americans who oppose, but what is left out is who conducted the poll, where it was conducted and whether that poll has changed more recently, as it was an October poll and this article was published in November.

    2. The construction also removed six historical trees and the Jacqueline Kennedy Garden to make way for a new ballroom.

      This line reinforces the narrative of destruction and harm and invokes loss of a historical building and nature.

    3. Satellite images show extent of demolition at White HouseNew before-and-after images show the former East Wing site, now completely demolished to make way for President Donald Trump’s ballroom, which is expected to cost about $300 million.

      In this print, it is an article that includes a visual reference, a news clip. This news clip includes satellite images of the white house over time that visualizes the destruction. In the caption in frames the story in a way that highlights the word demolition and frames it in a way that states it is President Trumps doing.

    4. New images show White House East Wing leveled for $300 million ballroom

      This headline constructs the story using the key words such as"leveled" which means "give a flat and even surface to", this verb was used to imply something was destroyed. Stating $300 million adds a level of wealth and extravagance.

    1. Although variations between the Septuagint and the Masoretic Text, revealed by Dead Sea Scrolls, have sparked debates about textual accuracy, the Greek translation was a cultural bridge that broadened the faith’s reach while subtly reshaping it.

      It’s interesting to learn that the Septuagint both spread the faith and subtly changed it, even with differences from the Masoretic Text. This makes me curious about how translations can shape a religion’s message.

    1. Media research methods are the practical procedures for carrying out a research project. These methods include content analysis, surveys, focus groups, experiments, and participant observation. Research methods generally involve either test subjects or analysis of media. Methods involving test subjects include surveys, depth interviews, focus groups, and experiments. Analysis of media can include content, style, format, social roles, and archival analysis.

      These forms of methodology can be very useful for learning how media influences culture and behavior.

    1. What was missing in this discourse was a recognition of the institutional policies and practices – includ-ing vastly unequal resources, a Eurocentric curriculum, teachers who were poorly prepared to teach students of diverse backgrounds and, of course, racism and other biases – that made educational inequality a natural outcome for large segments of the population.

      This passage helps name the shift from blaming children to fixing systems. Multicultural education was never just about adding a few new texts. It was a response to deficit stories that ignored unequal funding, a narrow Eurocentric curriculum, weak teacher preparation for diverse classrooms, and everyday racism. When we see those forces clearly, the work changes from trying to repair “culturally deprived” kids to redesigning schools with better resources, culturally sustaining content, and teachers trained to teach every student well. That is how you move the American dilemma from description to action.

    1. Other researchers admit that individuals prone to violent acts are indeed drawn to violent media; however, they claim that by keeping these individuals in a movie theater or at home, violent media have actually contributed to a reduction in violent social acts.

      It is important to understand the connection between violent media and real-world violent behavior. It makes sense that video games may be affecting people's emotions and thoughts, but there is likely another factor in a situation like the Columbine shooting.

    2. propaganda is not inherently good or bad. Whether propaganda has a positive or negative effect on society and culture depends on the motivations of those who use it and the understandings of those who receive it.

      Social media has been used by governments, organizations, and individuals as a platform to share and spread propaganda.

    1. He argues that recent research into the importance of hope for life outcomes is a “major break-through in thinking” for scholars in public health and epidemiology (p. 3). Syme attributes the genesis of this breakthrough to the groundbreaking White-hall studies, which led to revelations that the distribution of “virtually every

      Syme’s framing of hope as control of destiny really lands. It shifts the focus from pep talks to power. If unequal health tracks with class because control is uneven, schools can answer by giving students real agency over their learning and futures. That means clear and predictable systems, authentic choices in work, chances to revise, advisory that helps set and pursue goals, and concrete navigation help for college and aid. It also means reducing daily stressors with stable routines, mental health access, and help with basics so students can show up ready to learn. Do that and hope is not a mood. It is a capacity students practice every day.

    1. Simply adding Two-Spirit to the list of LGBTQ fails to fully account for the underlying systems ofpower and knowledge that continue to shape possibilities for solidarity be-tween queer and Indigenous communities.

      This relates to Combahee Collective's (2022) criticism of “add-and-stir” feminism that includes but does not transform systems. Inclusion that ignores power is not decolonization; it is reproducing existing colonial structures.

    2. The term is not intended to mark a singular identity category,but is an indigenously defined pan-Native North American term that refersto a diversity of Indigenous LGBTQ identities, as well as culturally specificnon-binary expressions of gender

      The discussion of Two-Spirit identity deepened my understanding of gender diversity as rooted in specific Indigenous worldviews; not just Western LGBTQ+ frameworks.

    3. We suggestthat this decolonization is already active in embodiments of Indigeneity andqueer gender and sexuality, yet this ongoing resistance to colonialism oftengoes unseen within queer spaces.

      This shows that the gender binary itself is a colonial project, reinforcing the gendered violence of settlement (Maracle,1996). The article shows how queer Indigenous and settler families resist these norms at home, which is often not recognized as important as public activism.

    4. Their stories, poetry, and cre-ative nonfiction constitute “theories for decolonizing the body and mind aswell a queer theory itself” (Burford, 2013, p. 168), yet their analysis and lead-ership are often marginalized within both LGBTQ and Indigenous scholar-ship, including in accounts of contemporary Two-Spirit community-building

      This parallels Yamada’s point on Asian American women’s invisibility (2022). It also mirrors the Combahee Collective’s claim that Black women are marginalized even within feminist spaces (2022). Silence from marginalized groups will not help protect them; to resist erasure one has to stand up and call out systems and practices of oppression. These texts highlight that marginalized queer people face erasure even inside supposedly inclusive movements.

    5. decolonization in our everyday lives, rather than only academic reflection onthese themes.

      This mirrors Anzaldúa and Moraga’s (2022) “theory in the flesh”: theory emerges from lived realities of race, sexuality, land, and oppression. Decolonization isn't only academic; it's embodied practice.

    6. it also canmask the reality that Cindy’s family benefits materially from owning land andthis ability to buy land is directly related to other forms of prior land theftof her White settler ancestors (and her partner’s

      There is material privilege from genocide. This reminds me of the statement "No one is illegal on stolen Indigenous lands", used to condemn the abuse that ICE is perpetrating in the US right now. Link to an article: https://www.culturalsurvival.org/news/no-one-illegal-stolen-indigenous-lands

    7. and decolo-nization that happen within and across intimate geographies of the home

      Article thesis that decolonization is lived through daily, small actions, not just grand public activism.

    8. as we redefine what it means to decolonize and to queer our homes, bodies,and relationships

      Who carries the emotional labour in these relationships; Indigenous people, settlers, or both?

    Annotators

    1. Ads might appeal to vanity, insecurity, prejudice, fear, or the desire for adventure.

      As I mentioned in a previous annotation, the constant stream of ads has affected many people's self-image. How should people feel if they are constantly shown people who don't look like them, trying to sell them a product that will supposedly make them happier or solve some problem? In our current technological landscape, people, or rather their data, have become the product.

    1. If culture is the expressed and shared values, attitudes, beliefs, and practices of a social group, organization, or institution, then what is popular culture? Popular culture is the media, products, and attitudes considered to be part of the mainstream of a given culture and the everyday life of common people. It is often distinct from more formal conceptions of culture that take into account moral, social, religious beliefs and values, such as our earlier definition of culture.

      Is popular culture actually distinct from moral, social, and religious beliefs? When all of those shape popular culture, the two can't truly be separated. For example, there are a multitude of books, songs, films, and other media that urge the importance and necessity of marriage. These ideas are ingrained into people's minds from a young age and will likely influence them as adults who will go on to create new pop culture media.

    1. Indian society into four classes or varna, priests and scholars, warriors and rulers, merchants and farmers, and laborers and servants.

      Learning about the four varnas shows me how Indian society was organized into clear social groups. It’s interesting to see how each group had its own role.

    1. Instead, accommodations have become another way for the most privileged students to press their advantage.

      This is an old, tired argument that people who know nothing about what it's like to have invisible disabilities have trotted out time and again since the ADA's passage in 1990.

    2. “Once they’re past Brown and off in the real world, that’s going to affect their performance.”

      EMPLOYERS ALSO HAVE TO PROVIDE ACCOMMODATIONS FOR FUCK'S SAKE

    3. Having ADHD or anxiety, for example, might make it difficult to focus. But focusing is a skill that the educational system is designed to test.

      YUP, THE AUTHOR DOESN'T KNOW ANYTHING ABOUT ADHD

    4. After the release of the DSM‑5, the symptoms needed only to “interfere with, or reduce the quality” of, academic functioning.

      this author is clearly dismissive of medical experts and probably ADHD.

    5. The change has occurred disproportionately at the most prestigious and expensive institutions.

      Yes, because people who attend those institutions tend to have access to health AND mental healthcare.

    1. Cultural imperialism can be a formal policy or can happen more subtly, as with the spread of outside influence through television, movies, and other cultural projects.

      This makes me think of how U.S. territories often us the names of U.S. presidents or states for roads names or currency. One could argue this is a form of cultural imperialism.

    1. Note: This preprint has been reviewed by subject experts for Review Commons. Content has not been altered except for formatting.

      Learn more at Review Commons


      Referee #3

      Evidence, reproducibility and clarity

      Summary

      The authors report the construction and validation of a novel SOX2-TBXT dual-reporter mouse embryonic stem cell (mESC) line, a tool enabling the simultaneous, live visualization of SOX2 and TBXT expression. Using this line, they established an in vitro differentiation system that generates populations of SOX2/TBXT co-expressing cells which mimic neuromesodermal progenitors (NMPs). The authors combined clonal analysis to assess the fate potential of these cells and applied gastruloid models to dissect the functional consequences of gene expression dynamics on axis elongation. The central finding is that the level of TBXT expression predicts and directs lineage potential. This work identifies specific expression thresholds that bias progenitors toward a mesodermal fate.

      Major comments

      The study's conclusions would be substantially strengthened by more direct validation in an embryonic context. Could the authors quantify endogenous SOX2 and TBXT protein levels of the NMPs in the mouse embryos with immunofluorescence? This would test whether a similar heterogeneity in TBXT expression exists in vivo and whether it correlates with the cells' spatial position.

      The role of SOX2 in the quantitative model seems underdeveloped. To justify the authors' claim, could the authors analyze their existing imaging data in more detail to disentangle if the absolute TBXT level is essential rather than the ratio of Sox2 to Tbxt is the driving factor for determining NMP fate?

      It seems that SOX2 expression also appears heterogeneous from both in vitro differentiation and gastruloid models. Quantifications, and a discussion of this heterogeneity and whether it influences the fate-decision process would be helpful.

      To support the authors' claim that the reporter line recapitulates endogenous protein expression in vivo (lines 121-128), please include a control immunostaining of wild-type embryo for SOX2 and TBXT to compare expression patterns side-by-side with those embryos shown in Figure 1B, Suppl. Fig. 1C, D. To substantiate the claims regarding cellular expression patterns within the embryo (line 125-128), the use of higher-resolution imaging, such as confocal microscopy, is recommended.

      The differentiation trajectory described culminates in a double-negative (Sox2-mCherry-negative, Tbxt-GFP-negative) population (lines 246-247). To provide a more complete picture of this fate progression, could the authors perform qPCR for relevant lineage markers to validate the molecular identity of this terminal population?

      Minor comments

      Scale bars for micrographs are missing in Figure 1B, Suppl. Fig. 1C, and D. The claims regarding the dynamics of TBXT and SOX2 expression in gastruloids following WNT/NOTCH inhibition (Figure 4B, 4D, 4E) would be more compelling if the authors include supplementary videos of the time-lapse imaging.

      In lines 322-324, the authors conclude that Tbxt-cells are the driving cells. Please elaborate on the interpretation that this is a cell-autonomous effect driven by TBXT levels. The observation that Sox2 levels increase ~10-15 hours after WNT/NOTCH inhibition is interesting (Figure 4D, 4E). Could the authors discuss this upregulation?

      Significance

      General Assessment

      In the development of a novel dual Sox2/Tbxt reporter cell line, which provides a powerful tool for quantitatively understanding the dynamics of cell fate specification during gastrulation and potentially in other developmental contexts. However, a key limitation is the study's primary focus on in vitro models. The findings will require further validation in an in vivo context.

      Advance

      This study provides a technical advance that provides a new resource available to the field for stem cell and developmental biology.

      Audience

      This paper will primarily interest a specialized audience, particularly developmental and stem cell biologists who study the fundamental mechanisms of embryogenesis, cell fate specification, and axis elongation.

      Field of expertise

      Stem cell biology and developmental biology. I do not have the expertise to evaluate their mathematical modeling.

    2. Note: This preprint has been reviewed by subject experts for Review Commons. Content has not been altered except for formatting.

      Learn more at Review Commons


      Referee #2

      Evidence, reproducibility and clarity

      Summary:

      In this manuscript, the authors generate a novel reporter mouse ESC line to track SOX2 and TBXT dynamics in neuromesodermal progenitors. The authors leverage multiple systems (in vitro differentiation, chimeric embryos and gastruloids) to address how SOX2/TBXT levels impact the contribution of neuromesodermal progenitors towards neural versus mesodermal fates. They show that the levels of TBXT can predict differentiation outcomes, whereby certain thresholds of TBXT influence the differentiation towards a mesodermal identity. In gastruloids, perturbing either WNT or NOTCH signalling coincides with diminished axial elongation. In vitro, the bias mediated by TBXT is also somewhat influenced by the substrate.

      Major comments:

      There are some key items to clarify that are important to resolve the interpretation of the results and clarify the main advances for a broader readership.

      With respect to the mESC differentiation into NMPs (using an established protocol from Gouti et al 2014), after Day 3, cells are cultured in conditions where they are exposed to both CHIR and FGF for a further two days. As these extended CHIR/FGF conditions don't appear to be characterised by Gouti et al., 2014, what proportions and progenitors are generated in the dish under these conditions? The loss of SOX2 by day 5 suggests mesodermal progenitors are the main derivatives but further characterization (eg Meox1/Msgn1) would be needed to verify this claim.

      Further validation of the generated derivatives would also be useful in the re-plating experiments (Figure 3) to test whether the double negative cells are transitioning to a mesodermal (eg Meox1/Msgn1) or neural derivative (eg Sox1/Pax6). Similarly, at day 5 and 6 of the differentiation, there appears to be a loss of Sox2 expression in some of the replated cells from the Sox2-positive population (see Figure 3D). Could the authors please clarify whether the double negative cells represent neural progenitors, and/or alternative cell types? Do the replated cells transiently adopt Tbxt? This would be possible by staining with neural (SOX1), or (pre)somitic mesoderm genes (MSGN1,MEOX1) or adjusting the text to reflect the uncertainty.

      At line 172 "The cells posterior to the node expressed only TBXT." Do these cells have low SOX2 expression that is hard to detect? Are these TBXT-positive cells derived from the primitive streak? Would staining with the primitive streak marker TBX6 enable visualization of these distinct cell types, and/or could the authors please label the figure in more detail.

      Could the authors please comment on the design of the reporter system in a bit more detail? For example, please clarify the necessity to generate a TBXT reporter that includes a H2B-GFP, unlike Sox2, which does not include a H2B. Can the authors distinguish between an increase in the threshold of TBXT, versus an increase detected due to the stability of the H2B-GFP? The low versus high TBXT cells may reflect early versus late TBXT expressing cells. Are the changes in TBXT expression (eg Supplementary Figure 2) significantly changed between the low versus high GFP populations? Additionally, why is the level of GFP similar across low and high GFP populations in Sup Fig 2? Could the in vivo data be used to quantify differences in TBXT (similar to what has been shown previously in Ivanovitch et al., 2021 PMID: 33999917)?

      Optional

      • Altering extrinsic cues (such as RA/CHIR) could clarify how reversible the high TBXT state is as cells progress towards mesoderm. Can you redirect the TBXT-high cells to a neural fate or are they already irreversibly committed?
      • Line 283: Flk1 is also expressed in TBXT-low cells. It would be interesting to test whether the TBXT-low cells and the SOX2neg/TBXTpos can generate lateral plate mesoderm or whether this competence to generate lateral plate mesoderm is limited to TBXT-high cells.
      • Line 324: The lack of elongation in gastruloids following the inhibition of WNT/NOTCH is clear. Do the authors expect that the reintroduction of TBXT would rescue axis elongation in the WNT/NOTCH inhibited gastruloids?

      Minor comments:

      Figure 3B - The SOX2+/TBXT- population only shows a moderate level of SOX1 by RT-qPCR. Using pre-neural markers (e.g. NKX1-2) might be helpful here to show progression towards a neural progenitor identity.

      Line 310 - Can you comment on the general efficiency in generating elongating gastruloids compared to WT cells and/or previous literature?

      Line 236 - Add "at constant SOX2 levels" (when comparing orange and yellow populations)

      Figure 5

      Fig5C and E take time to understand. Potentially expanding the figure legend slightly could be helpful to the reader.

      Supplementary Figure 2

      Line 864 - refer to Fig3A

      Line 350 - The link with testing the different substrates is a bit abrupt. Please can you make it clearer by modifying the text to explain the hypothesis being tested here.

      Line 363 - Could you comment in relation to what happens in the embryo to future mesoderm progenitors that seem to have a more motile phenotype compared to neural progenitors (eg Romanos et al 2021 PMID: 34607629)?

      Significance

      In this work, the authors have explored how the dynamics of SOX2/TBXT impact the decision process of neuromesodermal progenitors (NMPs). They have engineered and validated a novel dual fluorescent reporter ESC line to track SOX2 and TBXT. The work combines in vitro, in vivo and modelling approaches to understand how NMPs make decisions - a highly relevant and important question for multiple fields spanning developmental and quantitative biology, and the engineering of cell types in vitro from pluripotent cells. The authors propose a critically important finding: that discrete thresholds of TBXT influence the outcome of NMPs. However, further clarification is required to solidify these claims (discussed above).

      The data generated from this study also suggests that in contrast to Tbxt, the level of Sox2 does not appear to impact the NMP fate decision. While these are interesting and important findings, it is not entirely clear in this version of the manuscript how these advances relate with previous studies that have highlighted critical roles mediated by Sox2 and its level of expression (including the work of Koch et al 2017 - PMID: 28826820 and Blassberg et al PMID: 35550614). We expect that a broader discussion will in turn broaden the general interest and value of the work to a wider readership.

    3. Note: This preprint has been reviewed by subject experts for Review Commons. Content has not been altered except for formatting.

      Learn more at Review Commons


      Referee #1

      Evidence, reproducibility and clarity

      The work of Binagui-Casas, Granes and colleagues, investigates in a rigorous way the origin and potential of murine NMPs. The authors first validated a dual reporter system, which monitors Sox2 and Tbxt expression. Next, they identified the location in the embryo and the sequence of gene activation (Sox2 first, then Tbxt) leading to NMPs specification. Importantly, the in vitro model faithfully mimics the in vivo ontogeny of NMPs. Among Sox2+ NMCs, the authors observed different levels of Tbxt, which they proposed mark different stages of mesoderm maturation, with high Tbxt corresponding to more mature mesodermal state. Through sorting and replating of different populations, the authors convincingly showed that Sox2+/Tbxt-low cells are still bipotent, while Sox2+/Tbxt-high are committed towards the mesoderm lineage. Using a gastruloid model, the authors then showed that Tbxt expression correlates with axis elongation, as both are reduced upon inhibition of either WNT or NOTCH Finally, single-cell sorting followed by differentiation in FGF/CHIR and high throughput microscopy confirmed that double Sox2/Tbxt positive cells behave as NMPs and that high levels of Tbxt predispose cells towards mesoderm differentiation. These conclusions were further supported by mathematical modelling. The manuscript is easy to read, and figures are very clear. I have only minor suggestions, as I find the manuscript quite solid and complete.

      Minor points:

      1. The reporter system is based on stable fluorescent proteins, whose half-life is generally much longer than the endogenous proteins. This could generate a discrepancy between expression of reporters and the endogenous proteins. I find this relevant for Tbxt, as it is very clear that Tbxt reporter levels dictates the differentiation propensity, but I wonder whether, Tbxt low cells actually express TBXT protein or not. It might be the case that only a fraction of Tbxt low cells actually express TBXT protein, or none. It would enough to sort the populations showed in Figure 3A and perform immunostaining for endogenous SOX2 and TBXT. This could reveal even better correlations between their levels and cell behaviour.
      2. In the experiments in which IWP2 and LY411575 are used, I would suggest to asses cell viability, as the two inhibitors could induce toxicity. Staining for cleaved caspase-3 or a TUNEL assay would be enough. It would also be important to confirm that IWP2 blocks WNT signalling (by looking at WNT target genes or staining for active beta-cateinin) and that LY411575 blocks NOTCH signalling.
      3. I would define in the figure legends what the black line in figure 5E represents.

      Referees cross-commenting

      I agree with Reviewer #2 comment about "Further validation of the generated derivatives" by staining for additional markers.

      I also made a comment related to GFP stability, as Reviewer #2 did (i.e. The low versus high TBXT cells may reflect early versus late TBXT expressing cells ).

      Significance

      Overall, the manuscript uses an elegant approach and address an important question about NMPs behaviour. The results presented are an important advance in knowledge of NMP biology. I am confident that both stem cell and developmental biologist would be interested in this manuscript. I am an expert of pluripotency, signaling and models of early mammalian development.

    1. We hope that by the end of this book, you have a familiarity with applying different ethics frameworks, and considering the ethical tradeoffs of uses of social media and the design of social media systems.

      In another Informatics class I took, I encountered the question "Do social media and chatbots have an overall net positive effect on children?", and I felt like applying ethics principles such as consequentialism and taoism were useful in taking a stance on this question.

    2. We hope that by the end of this book, you have a familiarity with applying different ethics frameworks, and considering the ethical tradeoffs of uses of social media and the design of social media systems. Again, our goal has been not necessarily to come to the “right” answer, but to ask good questions and better understand the tradeoffs, unexpected side-effects, etc.

      After taking this course, I learned so many ethical frameworks, and I tried to utilized it when I faced a problem. I also slowly developed my own ethical framework.

    3. We hope that by the end of this book you know a lot of social media terminology (e.g., context collapse, parasocial relationships, the network effect, etc.), that you have a good overview of how social media works and is used, and what design decisions are made in how social media works, and the consequences of those decisions.

      During this quarter's lecture, I have come to understand the multifaceted nature of social media, from data and algorithms to harassment, mental health, and even capitalism and colonialism, all demonstrating its complexity. It made me realize that social media is not only a technical platform, but also supported by many backgrounds. After understanding these concepts, I have gained a clearer understanding of how the platform operates and how its decisions affect users and society.

    1. The medium was able to downplay regional differences and encourage a unified sense of the American lifestyle—a lifestyle that was increasingly driven and defined by consumer purchases. “Americans in the 1920s were the first to wear ready-made, exact-size clothing…to play electric phonographs, to use electric vacuum cleaners, to listen to commercial radio broadcasts, and to drink fresh orange juice year round.”Digital History, “The Formation of Modern American Mass Culture,” The Jazz Age: The American 1920s, 2007, www.digitalhistory.uh.edu/dat...y.cfm?hhid=454(accessed July 15, 2010). This boom in consumerism put its stamp on the 1920s, and, ironically, helped contribute to the Great Depression of the 1930s.

      This is true for certain demographics, but it can't be understated how consumer culture has become synonymous with American culture.

    2. An early advertising consultant claimed that the early days of radio were “a glorious opportunity for the advertising man to spread his sales propaganda” thanks to “a countless audience, sympathetic, pleasure seeking, enthusiastic, curious, interested, approachable in the privacy of their homes.

      This intrusion of advertisements in the home was only a precursor to our modern word, where ads are plastered over bus stops, grocery stores, entire buildings, and endless notification in our phones. I grew up in a word where I was always being advertised something, and it can be hard to imagine what it must have been like before.

    1. How have your views on social media changed (or been reinforced)?

      My views and how I interact with others on social media has changed in a positive way. I believe that I act more ethically conscious with the way I interact with others, considering their positions before saying anything and looking into how the content I consume is created before choosing to interact with it. For example, I do not support the usage of AI because it negatively effects the environment.

    2. How have your views on social media changed (or been reinforced)?

      I would say my views on social media are the same as when I came into this class, but they have been reinforced even more. My view on social media is a negative view. I think it is interesting and kind of cool that people across the entire world can share ideas and connect on the internet and social media. Human beings across the world have never been this connected, 100 years ago, even less actually, you would not have been able to communicate with someone in Europe from the United States so easily. We can do it within seconds nowadays and social media is one of the things that can create a collection of ideas from all different countries. That is what I enjoy about social media. But I believe it is made to be addictive and a lot of the time that people spend on social media these days is not productive or helpful. It can suck away the time in a day and make people lazy. It gives instant gratification and the companies behind social medias are not looking for the users best interest. All they want are to maximize the amount of time users' eyes are looking at their app. This view has not changed since reading this textbook and it had been even more reinforced.

    1. However, during the 19th century, several crucial inventions—the internal combustion engine, steam-powered ships, and railways, among others—led to other innovations across various industries. Suddenly, steam power and machine tools meant that production increased dramatically. But some of the biggest changes coming out of the Industrial Revolution were social in character. An economy based on manufacturing instead of agriculture meant that more people moved to cities, where techniques of mass production led to an emphasis on efficiency both in and out of the factory. Newly urbanized factory laborers no longer had the skill or time to produce their own food, clothing, or supplies and instead turned to consumer goods. Increased production led to increases in wealth, though income inequalities between classes also started to grow as well. Increased wealth and nonrural lifestyles led to the development of entertainment industries.

      Industrialization has forever changed our world. The advancements in technology and changes to labor that resulted from the industrial revolution changed the production and spread of mass media.

    1. Ever since Kennedy, American presidential hopefuls have had to be increasingly television-ready and media savvy. Indeed, evolving technology has helped change what the American public wants out of its leaders.

      Technology has changed what people expect of elected officials. Perhaps this is why many elected officials came to power from personality cults.

    2. So mass mediarefers to those means of transmission that are designed to reach a wide audience. Mass media are commonly considered to include radio, film, newspapers, magazines, books, and video games, as well as Internet blogs, podcasts, and video sharing.

      The webster definition of transmission is to "to send or convey from one person or place to another" or "to cause or allow to spread".

    1. All historians believe in honoring the integrity ofthe historical record. They do not fabricate evidence. Forgery and fraud violate the mostbasic foundations on which historians construct their interpretations of the past.

      Making up evidence destroys trust and damages the profession, and misleads the public. Because history relies on strong sources, any kind of dishonesty ruins the accuracy of the story being told. This quote also implies that integrity matters more than producing a certain argument or narrative.

    1. eLife Assessment

      This manuscript presents valuable findings which reveal an intricate pattern of memory expression following retrieval extinction at different intervals from retrieval-extinction to test. The novel advance is in the demonstration that, relative to a standard extinction procedure, the retrieval-extinction procedure more effectively suppresses responses to a conditioned threat stimulus when testing occurs just minutes after extinction. While the data provide solid evidence that the "short-term" suppression of responding involves engagement of the dorsolateral prefrontal cortex, there are inconsistencies in the analyses reported which obscure the interpretation and leave some of the claims with limited evidence.

    2. Reviewer #1 (Public review):

      Summary:

      The novel advance by Wang et al is in the demonstration that, relative to a standard extinction procedure, the retrieval-extinction procedure more effectively suppresses responses to a conditioned threat stimulus when testing occurs just minutes after extinction. The authors provide solid evidence to show that this "short-term" suppression of responding involves engagement of the dorsolateral prefrontal cortex.

      Strengths:

      Overall, the study is well-designed and the results are valuable. There are, however, a few issues in the way that it is introduced and discussed. It would have been useful if the authors could have more explicitly related the results to a theory - it would help the reader understand why the results should have come out the way that they did. More specific comments are presented below.

      Please note: The authors appear to have responded to my original review twice. It is not clear that they observed the public review that I edited after the first round of revisions. As part of these edits, I removed the entire section titled Clarifications, Elaborations and Edits

      Theory and Interpretation of Results

      (1) It is difficult to appreciate why the first trial of extinction in a standard protocol does NOT produce the retrieval-extinction effect. This applies to the present study as well as others that have purported to show a retrieval-extinction effect. The importance of this point comes through at several places in the paper. E.g., the two groups in study 1 experienced a different interval between the first and second CS extinction trials; and the results varied with this interval: a longer interval (10 min) ultimately resulted in less reinstatement of fear than a shorter interval. Even if the different pattern of results in these two groups was shown/known to imply two different processes, there is nothing in the present study that addresses what those processes might be. That is, while the authors talk about mechanisms of memory updating, there is little in the present study that permits any clear statement about mechanisms of memory. The references to a "short-term memory update" process do not help the reader to understand what is happening in the protocol.

      In reply to this point, the authors cite evidence to suggest that "an isolated presentation of the CS+ seems to be important in preventing the return of fear expression." They then note the following: "It has also been suggested that only when the old memory and new experience (through extinction) can be inferred to have been generated from the same underlying latent cause, the old memory can be successfully modified (Gershman et al., 2017). On the other hand, if the new experiences are believed to be generated by a different latent cause, then the old memory is less likely to be subject to modification. Therefore, the way the 1st and 2nd CS are temporally organized (retrieval-extinction or standard extinction) might affect how the latent cause is inferred and lead to different levels of fear expression from a theoretical perspective." This merely begs the question: why might an isolated presentation of the CS+ result in the subsequent extinction experiences being allocated to the same memory state as the initial conditioning experiences?<br /> This is not addressed in the paper. The study was not designed to address this question; and that the question did not need to be addressed for the set of results to be interesting. However, understanding how and why the retrieval-extinction protocol produces the effects that it does in the long-term test of fear expression would greatly inform our understanding of how and why the retrieval-extinction protocol has the effects that it does in the short-term tests of fear expression. To be clear; the results of the present study are very interesting - there is no denying that. I am not asking the authors to change anything in response to this point. It simply stands as a comment on the work that has been done in this paper and the area of research more generally.

      (2) The discussion of memory suppression is potentially interesting but raises many questions. That is, memory suppression is invoked to explain a particular pattern of results but I, as the reader, have no sense of why a fear memory would be better suppressed shortly after the retrieval-extinction protocol compared to the standard extinction protocol; and why this suppression is NOT specific to the cue that had been subjected to the retrieval-extinction protocol. I accept that the present study was not intended to examine aspects of memory suppression, and that it is a hypothesis proposed to explain the results collected in this study. I am not asking the authors to change anything in response to this point. Again, it simply stands as a comment on the work that has been done in this paper.

      (3) The authors have inserted the following text in the revised manuscript: "It should be noted that while our long-term amnesia results were consistent with the fear memory reconsolidation literatures, there were also studies that failed to observe fear prevention (Chalkia, Schroyens, et al., 2020; Chalkia, Van Oudenhove, et al., 2020; Schroyens et al., 2023). Although the memory reconsolidation framework provides a viable explanation for the long-term amnesia, more evidence is required to validate the presence of reconsolidation, especially at the neurobiological level (Elsey et al., 2018). While it is beyond the scope of the current study to discuss the discrepancies between these studies, one possibility to reconcile these results concerns the procedure for the retrieval-extinction training. It has been shown that the eligibility for old memory to be updated is contingent on whether the old memory and new observations can be inferred to have been generated by the same latent cause (Gershman et al., 2017; Gershman and Niv, 2012). For example, prevention of the return of fear memory can be achieved through gradual extinction paradigm, which is thought to reduce the size of prediction errors to inhibit the formation of new latent causes (Gershman, Jones, et al., 2013). Therefore, the effectiveness of the retrieval-extinction paradigm might depend on the reliability of such paradigm in inferring the same underlying latent cause." ***It is perfectly fine to state that "the effectiveness of the retrieval-extinction paradigm might depend on the reliability of such paradigm in inferring the same underlying latent cause..." This is not uninteresting; but it also isn't saying much. Ideally, the authors would have included some statement about factors that are likely to determine whether one is or isn't likely to see a retrieval-extinction effect, grounded in terms of the latent state theories that have been invoked here. Presumably, the retrieval-extinction protocol has variable effects because of procedural differences that affect whether subjects infer the same underlying latent cause when shifted into extinction. Surely, the clinical implications of any findings are seriously curtailed unless one understands when a protocol is likely to produce an effect; and why the effect occurs at all? This question is rhetorical. I am not asking the authors to change anything in response to this point. Again, it stands as a comment on the work that has been done in this paper; and remains a comment after insertion of the new text, which is acknowledged and appreciated.

      (4) The authors find different patterns of responses to CS1 and CS2 when they were tested 30 min after extinction versus 24 h after extinction. On this basis, they infer distinct memory update mechanisms. However, I still can't quite see why the different patterns of responses at these two time points after extinction need to be taken to infer different memory update mechanisms. That is, the different patterns of responses at the two time points could be indicative of the same "memory update mechanism" in the sense that the retrieval-extinction procedure induces a short-term memory suppression that serves as the basis for the longer-term memory suppression (i.e., the reconsolidation effect). My pushback on this point is based on the notion of what constitutes a memory update mechanism; and is motivated by what I take to be a rather loose use of language/terminology in the reconsolidation literature and this paper specifically (for examples, see the title of the paper and line 2 of the abstract).

      To be clear: I accept the authors' reply that "The focus of the current manuscript is to demonstrate that the retrieval-extinction paradigm can also facilitate a short-term fear memory deficit measured by SCR". However, I disagree with the claim that any short-term fear memory deficit must be indicative of "update mechanisms other than reconsolidation", which appears on Line 27 in the abstract and very much indicates the spirit of the paper. To make the point: the present study has examined the effectiveness of a retrieval-extinction procedure in suppressing fear responses 30 min, 6 hours and 24 hours after extinction. There are differences across the time points in terms of the level of suppression, its cue specificity, and its sensitivity to manipulation of activity in the dlPFC. This is perfectly interesting when not loaded with additional baggage re separable mechanisms of memory updating at the short and long time points: there is simply no evidence in this study or anywhere else that the short-term deficit in suppression of fear responses has anything whatsoever to do with memory updating. It can be exactly what is implied by the description: a short-term deficit in the suppression of fear responses. Again, this stands as a comment on the work that has been done; and remains a comment for the revised paper.

      (5) It is not clear why thought control ability ought to relate to any aspect of the suppression that was evident in the 30 min tests - that is, I accept the correlation between thought control ability and performance in the 30 min tests but would have liked to know why this was looked at in the first place and what, if anything, it means. The issue at hand is that, as best as I can tell, there is no theory to which the result from the short- and long-term tests can be related. The attempts to fill this gap with reference to phenomena like retrieval-induced forgetting are appreciated but raise more questions than answers. This is especially clear in the discussion, where it is acknowledged/stated: "Inspired by the similarities between our results and suppression-induced declarative memory amnesia (Gagnepain et al., 2017), we speculate that the retrieval-extinction procedure might facilitate a spontaneous memory suppression process and thus yield a short-term amnesia effect. Accordingly, the activated fear memory induced by the retrieval cue would be subjected to an automatic fear memory suppression through the extinction training (Anderson and Floresco, 2022)." There is nothing in the subsequent discussion to say why this should have been the case other than the similarity between results obtained in the present study and those in the literature on retrieval induced forgetting, where the nature of the testing is quite different. Again, this is simply a comment on the work that has been done - no change is required for the revised paper.

    3. Reviewer #2 (Public review):

      Summary

      The study investigated whether memory retrieval followed soon by extinction training results in a short-term memory deficit when tested - with a reinstatement test that results in recovery from extinction - soon after extinction training. Experiment 1 documents this phenomenon using a between-subjects design. Experiment 2 used a within-subject control and sees that the effect is also observed in a control condition. In addition, it also revealed that if testing is conducted 6 hours after extinction, there is not effect of retrieval prior to extinction as there is recovery from extinction independently of retrieval prior to extinction. A third Group also revealed that retrieval followed by extinction attenuates reinstatement when the test is conducted 24 hours later, consistent with previous literature. Finally, Experiment 3 used continuous theta-burst stimulation of the dorsolateral prefrontal cortex and assessed whether inhibition of that region (vs a control region) reversed the short-term effect revealed in Experiments 1 and 2. The results of control groups in Experiment 3 replicated the previous findings (short-term effect), and the experimental group revealed that these can be reversed by inhibition of the dorsolateral prefrontal cortex.

      Strengths

      The work is performed using standard procedures (fear conditioning and continuous theta-burst stimulation) and there is some justification of the sample sizes. The results replicate previous findings - some of which have been difficult to replicate and this needs to be acknowledged - and suggest that the effect can also be observed in a short-term reinstatement test.

      The study establishes links between the memory reconsolidation and retrieval-induced forgetting (or memory suppression) literatures. The explanations that have been developed for these are distinct and the current results integrate these, by revealing that the DLPFC activity involved in retrieval-extinction short-term effect. There is thus some novelty in the present results, but numerous questions remain unaddressed.

      Weakness

      The fear acquisition data is converted to a differential fear SCR and this is what is analysed (early vs late). However, the figure shows the raw SCR values for CS+ and CS- and therefore it is unclear whether acquisition was successful (despite there being an "early" vs "late" effect - no descriptives are provided).

      In Experiment 1 (Test results) it is unclear whether the main conclusion stems from a comparison of the test data relative to the last extinction trial ("we defined the fear recovery index as the SCR difference between the first test trial and the last extinction trial for a specific CS") or the difference relative to the CS- ("differential fear recovery index between CS+ and CS-"). It would help the reader assess the data if Fig 1e presents all the indexes (both CS+ and CS-). In addition, there is one sentence which I could not understand "there is no statistical difference between the differential fear recovery indexes between CS+ in the reminder and no reminder groups (P=0.048)". The p value suggests that there is a difference, yet it is not clear what is being compared here. Critically, any index taken as a difference relative to the CS- can indicate recovery of fear to the CS+ or absence of discrimination relative to the CS-, so ideally the authors would want to directly compare responses to the CS+ in the reminder and no-reminder groups. In the absence of such comparison, little can be concluded, in particular if SCR CS- data is different between groups. The latter issue is particularly relevant in Experiment 2, in which the CS- seems to vary between groups during the test and this can obscure the interpretation of the result.

      In experiment 1, the findings suggest that there is a benefit of retrieval followed by extinction in a short-term reinstatement test. In Experiment 2, the same effect is observed to a cue which did not undergo retrieval before extinction (CS2+), a result that is interpreted as resulting from cue-independence, rather than a failure to replicate in a within-subjects design the observations of Experiment 1 (between-subjects). Although retrieval-induced forgetting is cue-independent (the effect on items that are supressed [Rp-] can be observed with an independent probe), it is not clear that the current findings are similar, and thus that the strong parallels made are not warranted. Here, both cues have been extinguished and therefore been equally exposed during the critical stage.

      The findings in Experiment 2 suggest that the amnesia reported in experiment 1 is transient, in that no effect is observed when the test is delayed by 6 hours. The phenomena whereby reactivated memories transition to extinguished memories as a function of the amount of exposure (or number of trials) is completely different from the phenomena observed here. In the former, the manipulation has to do with the number of trials (or total amount of time) that the cues are exposed. In the current Experiment 2, the authors did not manipulate the number of trials but instead the retention interval between extinction and test. The finding reported here is closer to a "Kamin effect", that is the forgetting of learned information which is observed with intervals of intermediate length (Baum, 1968). Because the Kamin effect has been inferred to result from retrieval failure, it is unclear how this can be explained here. There needs to be much more clarity on the explanations to substantiate the conclusions.

      There are many results (Ryan et al., 2015) that challenge the framework that the authors base their predictions on (consolidation and reconsolidation theory), therefore these need to be acknowledged. These studies showed that memory can be expressed in the absence of the biological machinery thought to be needed for memory performance. The authors should be careful about statements such as "eliminate fear memores" for which there is little evidence.

      The parallels between the current findings and the memory suppression literature are speculated in the general discussion, and there is the conclusion that "the retrieval-extinction procedure might facilitate a spontaneous memory suppression process". Because one of the basic tenets of the memory suppression literature is that it reflects an "active suppression" process, there is no reason to believe that in the current paradigm the same phenomenon is in place, but instead it is "automatic". In other words, the conclusions make strong parallels with the memory suppression (and cognitive control) literature, yet the phenomena that they observed is thought to be passive (or spontaneous/automatic). Ultimately, it is unclear why 10 mins between the reminder and extinction learning will "automatically" supress fear memories. Further down in the discussion it is argued that "For example, in the well-known retrieval-induced forgetting (RIF) phenomenon, the recall of a stored memory can impair the retention of related long-term memory and this forgetting effect emerges as early as 20 minutes after the retrieval procedure, suggesting memory suppression or inhibition can occur in a more spontaneous and automatic manner". I did not follow with the time delay between manipulation and test (20 mins) would speak about whether the process is controlled or automatic. In addition, the links with the "latent cause" theoretical framework are weak if any. There is little reason to believe that one extinction trial, separated by 10 mins from the rest of extinction trials, may lead participants to learn that extinction and acquisition have been generated by the same latent cause.

      Among the many conclusions, one is that the current study uncovers the "mechanism" underlying the short-term effects of retrieval-extinction. There is little in the current report that uncovers the mechanism, even in the most psychological sense of the mechanism, so this needs to be clarified. The same applies to the use of "adaptive".

      Whilst I could access the data in the OFS site, I could not make sense of the Matlab files as there is no signposting indicating what data is being shown in the files. Thus, as it stands, there is no way of independently replicating the analyses reported.<br /> The supplemental material shows figures with all participants, but only some statistical analyses are provided, and sometimes these are different from those reported in the main manuscript. For example, the test data in Experiment 1 is analysed with a two-way ANOVA with main effects of group (reminder vs no-reminder) and time (last trial of extinction vs first trial of test) in the main report. The analyses with all participants in the sup mat used a mixed two-way ANOVA with group (reminder vs no reminder) and CS (CS+ vs CS-). This makes it difficult to assess the robustness of the results when including all participants. In addition, in the supplementary materials there are no figures and analyses for Experiment 3.

      One of the overarching conclusions is that the "mechanisms" underlying reconsolidation (long term) and memory suppression (short term) phenomena are distinct, but memory suppression phenomena can also be observed after a 7-day retention interval (Storm et al., 2012), which then questions the conclusions achieved by the current study.

      References:

      Baum, M. (1968). Reversal learning of an avoidance response and the Kamin effect. Journal of Comparative and Physiological Psychology, 66(2), 495.<br /> Chalkia, A., Schroyens, N., Leng, L., Vanhasbroeck, N., Zenses, A. K., Van Oudenhove, L., & Beckers, T. (2020). No persistent attenuation of fear memories in humans: A registered replication of the reactivation-extinction effect. Cortex, 129, 496-509.<br /> Ryan, T. J., Roy, D. S., Pignatelli, M., Arons, A., & Tonegawa, S. (2015). Engram cells retain memory under retrograde amnesia. Science, 348(6238), 1007-1013.<br /> Storm, B. C., Bjork, E. L., & Bjork, R. A. (2012). On the durability of retrieval-induced forgetting. Journal of Cognitive Psychology, 24(5), 617-629.

      Comments on revisions:

      Thanks to the authors for trying to address my concerns.

      (1 and 2) My point about evidence for learning relates to the fact that in none of the experiments an increase in SCR to the CSs+ is observed during training (in Experiment 1 CS+/CS- differences are even present from the outset), instead what happens is that participants learn to discriminate between the CS+ and CS- and decrease their SCR responding to the safe CS-. This begs the question as to what is being learned, given that the assumption is that the retrieval-extinction treatment is concerned with the excitatory memory (CS+) rather than the CS+/CS- discrimination. For example, Figures 6A and 6B have short/Long term amnesia in the right axes, but it is unclear from the data what memory is being targeted. In Figure 6C, the right panels depicting Suppression and Reconsolidation mechanisms suggest that it is the CS+ memory that is being targeted. Because the dependent measure (differential SCR) captures how well the discrimination was learned (this point relates to point 2 which the authors now acknowledge that there are differences between groups in responding to the CS-), then I struggle to see how the data supports these CS+ conclusions. The fact that influential papers have used this dependent measure (i.e., differential SCR) does not undermine the point that differences between groups at test are driven by differences in responding to the CS-.

      (3, 4 and 5) The authors have qualified some of the statements, yet I fail to see some of these parallels. Much of the discussion is speculative and ultimately left for future research to address.

      (6) I can now make more sense of the publicly available data, although the files would benefit from an additional column that distinguishes between participants that were included in the final analyses (passed the multiple criteria = 1) and those who did not (did not pass the criteria = 0). Otherwise, anyone who wants to replicate these analyses needs to decipher the multiple inclusion criteria and apply it to the dataset.

    4. Author response:

      The following is the authors’ response to the previous reviews

      Reviewer #1 (Public review):

      Introduction & Theory

      (1) It is difficult to appreciate why the first trial of extinction in a standard protocol does NOT produce the retrieval-extinction effect. This applies to the present study as well as others that have purported to show a retrieval-extinction effect. The importance of this point comes through at several places in the paper. E.g., the two groups in Study 1 experienced a different interval between the first and second CS extinction trials; and the results varied with this interval: a longer interval (10 min) ultimately resulted in less reinstatement of fear than a shorter interval. Even if the different pattern of results in these two groups was shown/known to imply two different processes, there is nothing in the present study that addresses what those processes might be. That is, while the authors talk about mechanisms of memory updating, there is little in the present study that permits any clear statement about mechanisms of memory. The references to a "short-term memory update" process do not help the reader to understand what is happening in the protocol.

      We agree with the reviewer that whether and how the retrieval-extinction paradigm works is still under debate. Our results provide another line of evidence that such a paradigm is effective in producing long term fear amnesia. The focus of the current manuscript is to demonstrate that the retrieval-extinction paradigm can also facilitate a short-term fear memory deficit measured by SCR. Our TMS study provided some preliminary evidence in terms of the brain mechanisms involved in the causal relationship between the dorsolateral prefrontal cortex (dlPFC) activity and the short-term fear amnesia and showed that both the retrieval interval and the intact dlPFC activity were necessary for the short-term fear memory deficit and accordingly were referred to as the “mechanism” for memory update. We acknowledge that the term “mechanism” might have different connotations for different researchers. We now more explicitly clarify what we mean by “mechanisms” in the manuscript (line 99) as follows:

      “In theory, different cognitive mechanisms underlying specific fear memory deficits, therefore, can be inferred based on the difference between memory deficits.”

      In reply to this point, the authors cite evidence to suggest that "an isolated presentation of the CS+ seems to be important in preventing the return of fear expression." They then note the following: "It has also been suggested that only when the old memory and new experience (through extinction) can be inferred to have been generated from the same underlying latent cause, the old memory can be successfully modified (Gershman et al., 2017). On the other hand, if the new experiences are believed to be generated by a different latent cause, then the old memory is less likely to be subject to modification. Therefore, the way the 1stand 2ndCS are temporally organized (retrieval-extinction or standard extinction) might affect how the latent cause is inferred and lead to different levels of fear expression from a theoretical perspective." This merely begs the question: why might an isolated presentation of the CS+ result in the subsequent extinction experiences being allocated to the same memory state as the initial conditioning experiences? This is not yet addressed in any way.

      As in our previous response, this manuscript is not about investigating the cognitive mechanism why and how an isolated presentation of the CS+ would suppress fear expression in the long term. As the reviewer is aware, and as we have addressed in our previous response letters, both the positive and negative evidence abounds as to whether the retrieval-extinction paradigm can successfully suppress the long-term fear expression. Previous research depicted mechanisms instigated by the single CS+ retrieval at the molecular, cellular, and systems levels, as well as through cognitive processes in humans. In the current manuscript, we simply set out to test that in addition to the long-term fear amnesia, whether the retrieval-extinction paradigm can also affect subjects’ short-term fear memory.

      (2) The discussion of memory suppression is potentially interesting but, in its present form, raises more questions than it answers. That is, memory suppression is invoked to explain a particular pattern of results but I, as the reader, have no sense of why a fear memory would be better suppressed shortly after the retrieval-extinction protocol compared to the standard extinction protocol; and why this suppression is NOT specific to the cue that had been subjected to the retrieval-extinction protocol.

      Memory suppression is the hypothesis we proposed that might be able to explain the results we obtained in the experiments. We discussed the possibility of memory suppression and listed the reasons why such a mechanism might be at work. As we mentioned in the manuscript, our findings are consistent with the memory suppression mechanism on at least two aspects: 1) cue-independence and 2) thought-control ability dependence. We agree that the questions raised by the reviewer are interesting but to answer these questions would require a series of further experiments to disentangle all the various variables and conceptual questions about the purpose of a phenomenon, which we are afraid is out of the scope of the current manuscript. We refer the reviewer to the discussion section where memory suppression might be the potential mechanism for the short-term amnesia we observed (lines 562-569) as follows:

      “Previous studies indicate that a suppression mechanism can be characterized by three distinct features: first, the memory suppression effect tends to emerge early, usually 10-30 mins after memory suppression practice and can be transient (MacLeod and Macrae, 2001; Saunders and MacLeod, 2002); second, the memory suppression practice seems to directly act upon the unwanted memory itself (Levy and Anderson, 2002), such that the presentation of other cues originally associated with the unwanted memory also fails in memory recall (cue-independence); third, the magnitude of memory suppression effects is associated with individual difference in control abilities over intrusive thoughts (Küpper et al., 2014).”

      (3) Relatedly, how does the retrieval-induced forgetting (which is referred to at various points throughout the paper) relate to the retrieval-extinction effect? The appeal to retrieval-induced forgetting as an apparent justification for aspects of the present study reinforces points 2 and 3 above. It is not uninteresting but lacks clarification/elaboration and, therefore, its relevance appears superficial at best.

      We brought the topic of retrieval-induced forgetting (RIF) to stress the point that memory suppression can be unconscious. In a standard RIF paradigm, unlike the think/no-think paradigm, subjects are not explicitly told to suppress the non-target memories. However, to successfully retrieve the target memory, the cognitive system actively inhibits the non-target memories, effectively implementing a memory suppression mechanism (though unconsciously). Therefore, it is possible our results might be explained by the memory suppression framework. We elaborated this point in the discussion section (lines 578-584): 

      “In our experiments, subjects were not explicitly instructed to suppress their fear expression, yet the retrieval-extinction training significantly decreased short-term fear expression. These results are consistent with the short-term amnesia induced with the more explicit suppression intervention (Anderson et al., 1994; Kindt and Soeter, 2018; Speer et al., 2021; Wang et al., 2021; Wells and Davies, 1994). It is worth noting that although consciously repelling unwanted memory is a standard approach in memory suppression paradigm, it is possible that the engagement of the suppression mechanism can be unconscious.”

      (4) I am glad that the authors have acknowledged the papers by Chalkia, van Oudenhove & Beckers (2020) and Chalkia et al (2020), which failed to replicate the effects of retrieval-extinction reported by Schiller et al in Reference 6. The authors have inserted the following text in the revised manuscript: "It should be noted that while our long-term amnesia results were consistent with the fear memory reconsolidation literature, there were also studies that failed to observe fear prevention (Chalkia, Schroyens, et al., 2020; Chalkia, Van Oudenhove, et al., 2020; Schroyens et al., 2023). Although the memory reconsolidation framework provides a viable explanation for the long-term amnesia, more evidence is required to validate the presence of reconsolidation, especially at the neurobiological level (Elsey et al., 2018). While it is beyond the scope of the current study to discuss the discrepancies between these studies, one possibility to reconcile these results concerns the procedure for the retrieval-extinction training. It has been shown that the eligibility for old memory to be updated is contingent on whether the old memory and new observations can be inferred to have been generated by the same latent cause (Gershman et al., 2017; Gershman and Niv, 2012). For example, prevention of the return of fear memory can be achieved through gradual extinction paradigm, which is thought to reduce the size of prediction errors to inhibit the formation of new latent causes (Gershman, Jones, et al., 2013). Therefore, the effectiveness of the retrieval-extinction paradigm might depend on the reliability of such paradigm in inferring the same underlying latent cause." Firstly, if it is beyond the scope of the present study to discuss the discrepancies between the present and past results, it is surely beyond the scope of the study to make any sort of reference to clinical implications!!!

      As we have clearly stated in our manuscript that this paper was not about discussing why some literature was or was not able to replicate the retrieval-extinction results originally reported by Schiller et al. 2010. Instead, we aimed to report a novel short-term fear amnesia through the retrieval-extinction paradigm, above and beyond the long-term amnesia reported before. Speculating about clinical implications of these finding is unrelated to the long-term, amnesia debate in the reconsolidation world. We now refer the reader to several perspectives and reviews that have proposed ways to resolve these discrepancies as follows (lines 642-673).

      Secondly, it is perfectly fine to state that "the effectiveness of the retrieval-extinction paradigm might depend on the reliability of such paradigm in inferring the same underlying latent cause..." This is not uninteresting, but it also isn't saying much. Minimally, I would expect some statement about factors that are likely to determine whether one is or isn't likely to see a retrieval-extinction effect, grounded in terms of this theory.

      Again, as we have responded many times, we simply do not know why some studies were able to suppress the fear expression using the retrieval-extinction paradigm and other studies weren’t. This is still an unresolved issue that the field is actively engaging with, and we now refer the reader to several papers dealing with this issue. However, this is NOT the focus of our manuscript. Having a healthy debate does not mean that every study using the retrieval-extinction paradigm must address the long-standing question of why the retrieval-extinction paradigm is effective (at least in some studies).

      Clarifications, Elaborations, Edits

      (5) Some parts of the paper are not easy to follow. Here are a few examples (though there are others):

      (a) In the abstract, the authors ask "whether memory retrieval facilitates update mechanisms other than memory reconsolidation"... but it is never made clear how memory retrieval could or should "facilitate" a memory update mechanism.

      We meant to state that the retrieval-extinction paradigm might have effects on fear memory, above and beyond the purported memory reconsolidation effect. Sentence modified (lines 25-26) as follows:

      “Memory reactivation renders consolidated memory fragile and thereby opens the window for memory updates, such as memory reconsolidation.”

      (b) The authors state the following: "Furthermore, memory reactivation also triggers fear memory reconsolidation and produces cue specific amnesia at a longer and separable timescale (Study 2, N = 79 adults)." Importantly, in study 2, the retrieval-extinction protocol produced a cue-specific disruption in responding when testing occurred 24 hours after the end of extinction. This result is interesting but cannot be easily inferred from the statement that begins "Furthermore..." That is, the results should be described in terms of the combined effects of retrieval and extinction, not in terms of memory reactivation alone; and the statement about memory reconsolidation is unnecessary. One can simply state that the retrieval-extinction protocol produced a cue-specific disruption in responding when testing occurred 24 hours after the end of extinction.

      The sentence the reviewer referred to was in our original manuscript submission but had since been modified based on the reviewer’s comments from last round of revision. Please see the abstract (lines 30-35) of our revised manuscript from last round of revision:

      “Furthermore, across different timescales, the memory retrieval-extinction paradigm triggers distinct types of fear amnesia in terms of cue-specificity and cognitive control dependence, suggesting that the short-term fear amnesia might be caused by different mechanisms from the cue-specific amnesia at a longer and separable timescale (Study 2, N = 79 adults).”

      (c) The authors also state that: "The temporal scale and cue-specificity results of the short-term fear amnesia are clearly dissociable from the amnesia related to memory reconsolidation, and suggest that memory retrieval and extinction training trigger distinct underlying memory update mechanisms." ***The pattern of results when testing occurred just minutes after the retrieval-extinction protocol was different to that obtained when testing occurred 24 hours after the protocol. Describing this in terms of temporal scale is unnecessary; and suggesting that memory retrieval and extinction trigger different memory update mechanisms is not obviously warranted. The results of interest are due to the combined effects of retrieval+extinction and there is no sense in which different memory update mechanisms should be identified with the different pattern of results obtained when testing occurred either 30 min or 24 hours after the retrieval-extinction protocol (at least, not the specific pattern of results obtained here).

      Again, we are afraid that the reviewer referred to the abstract in the original manuscript submission, instead of the revised abstract we submitted in the last round. Please see lines 37-39 of the revised abstract where the sentence was already modified (or the abstract from last round of revision).

      The facts that the 30min, 6hr and 24hr test results are different in terms of their cue-specificity and thought-control ability dependence are, to us, an important discovery in terms of delineating different cognitive processes at work following the retrieval-extinction paradigm. We want to emphasize that the fear memories after going through the retrieval-extinction paradigm showed interesting temporal dynamics in terms of their magnitudes, cue-specificity and thought-control ability dependence.

      (d) The authors state that: "We hypothesize that the labile state triggered by the memory retrieval may facilitate different memory update mechanisms following extinction training, and these mechanisms can be further disentangled through the lens of temporal dynamics and cue-specificities." *** The first part of the sentence is confusing around usage of the term "facilitate"; and the second part of the sentence that references a "lens of temporal dynamics and cue-specificities" is mysterious. Indeed, as all rats received the same retrieval-extinction exposures in Study 2, it is not clear how or why any differences between the groups are attributed to "different memory update mechanisms following extinction"

      The term “facilitate” was used to highlight the fact that the short-term fear amnesia effect is also memory retrieval dependent, as study 1 demonstrated. The novelty of the short-term fear memory deficit can be distinguished from the long-term memory effect via cue-specificity and thought-control ability dependence. Sentence has been modified (lines 97-101) as follows:

      “We hypothesize that the labile state triggered by the memory retrieval may facilitate different memory deficits following extinction training, and these deficits can be further disentangled through the lens of temporal dynamics and cue-specificities. In theory, different cognitive mechanisms underlying specific fear memory deficits, therefore, can be inferred based on the difference between memory deficits.”

      Data

      (6A) The eight participants who were discontinued after Day 1 in Study 1 were all from the no reminder group. The authors should clarify how participants were allocated to the two groups in this experiment so that the reader can better understand why the distribution of non-responders was non-random (as it appears to be).

      (6B) Similarly, in study 2, of the 37 participants that were discontinued after Day 2, 19 were from Group 30 min and 5 were from Group 6 hours. The authors should comment on how likely these numbers are to have been by chance alone. I presume that they reflect something about the way that participants were allocated to groups: e.g., the different groups of participants in studies 1 and 2 could have been run at quite different times (as opposed to concurrently). If this was done, why was it done? I can't see why the study should have been conducted in this fashion - this is for myriad reasons, including the authors' concerns re SCRs and their seasonal variations.

      As we responded in the previous response letters (as well as in the revised the manuscript), subjects were excluded because their SCR did not reach the threshold of 0.02 S when electric shock was applied. Subjects were assigned to different treatments daily (eg. Day 1 for the reminder group and Day 2 for no-reminder group) to avoid potential confusion in switching protocols to different subjects within the same day. We suspect that the non-responders might be related to the body thermal conditions caused by the lack of central heating for specific dates. Please note that the discontinued subjects (non-responders) were let go immediately after the failure to detect their SCR (< 0.02 S) on Day 1 and never invited back on Day 2, so it’s possible that the discontinued subjects were all from certain dates on which the body thermal conditions were not ideal for SCR collection. Despite the number of excluded subjects, we verified the short-term fear amnesia effect in three separate studies, which to us should serve as strong evidence in terms of the validity of the effect.

      (6C) In study 2, why is responding to the CS- so high on the first test trial in Group 30 min? Is the change in responding to the CS- from the last extinction trial to the first test trial different across the three groups in this study? Inspection of the figure suggests that it is higher in Group 30 min relative to Groups 6 hours and 24 hours. If this is confirmed by the analysis, it has implications for the fear recovery index which is partly based on responses to the CS-. If not for differences in the CS- responses, Groups 30 min and 6 hours are otherwise identical. That is, the claim of differential recovery to the CS1 and CS2 across time may simply an artefact of the way that the recovery index was calculated. This is unfortunate but also an important feature of the data given the way in which the fear recovery index was calculated.

      We have provided detailed analysis to this question in our previous response letter, and we are posting our previous response there:

      Following the reviewer’s comments, we went back and calculated the mean SCR difference of CS- between the first test trial and the last extinction trial for all three studies (see Author response image 1 below). In study 1, there was no difference in the mean CS- SCR (between the first test trial and last extinction trial) between the reminder and no-reminder groups (Kruskal-Wallis test , though both groups showed significant fear recovery even in the CS- condition (Wilcoxon signed rank test, reminder: P = 0.0043, no-reminder: P = 0.0037). Next, we examined the mean SCR for CS- for the 30min, 6h and 24h groups in study 2 and found that there was indeed a group difference (one-way ANOVA,F<sub>2.76</sub> = 5.3462, P = 0.0067, panel b), suggesting that the CS- related SCR was influenced by the test time (30min, 6h or 24h). We also tested the CS- related SCR for the 4 groups in study 3 (where test was conducted 1 hour after the retrieval-extinction training) and found that across TMS stimulation types (PFC vs. VER) and reminder types (reminder vs. no-reminder) the ANOVA analysis did not yield main effect of TMS stimulation type (F<sub>1.71</sub> = 0.322, P = 0.572) nor main effect of reminder type (F<sub>1.71</sub> = 0.0499, P = 0.824, panel c). We added the R-VER group results in study 3 (see panel c) to panel b and plotted the CS- SCR difference across 4 different test time points and found that CS- SCR decreased as the test-extinction delay increased (Jonckheere-Terpstra test, P = 0.00028). These results suggest a natural “forgetting” tendency for CS- related SCR and highlight the importance of having the CS- as a control condition to which the CS+ related SCR was compared with.

      Author response image 1.

      (6D) The 6 hour group was clearly tested at a different time of day compared to the 30 min and 24 hour groups. This could have influenced the SCRs in this group and, thereby, contributed to the pattern of results obtained.

      Again, we answered this question in our previous response. Please see the following for our previous response:

      For the 30min and 24h groups, the test phase can be arranged in the morning, in the afternoon or at night. However, for the 6h group, the test phase was inevitably in the afternoon or at night since we wanted to exclude the potential influence of night sleep on the expression of fear memory (see Author response table 1 below). If we restricted the test time in the afternoon or at night for all three groups, then the timing of their extinction training was not matched.

      Author response table 1.

      Nevertheless, we also went back and examined the data for the subjects only tested in the afternoon or at nights in the 30min and 24h groups to match with the 6h group where all the subjects were tested either in the afternoon or at night. According to the table above, we have 17 subjects for the 30min group (9+8),18 subjects for the 24h group (9 + 9) and 26 subjects for the 6h group (12 + 14). As Author response image 2 shows, the SCR patterns in the fear acquisition, extinction and test phases were similar to the results presented in the original figure.

      Author response image 2.

      (6E) The authors find different patterns of responses to CS1 and CS2 when they were tested 30 min after extinction versus 24 h after extinction. On this basis, they infer distinct memory update mechanisms. However, I still can't quite see why the different patterns of responses at these two time points after extinction need to be taken to infer different memory update mechanisms. That is, the different patterns of responses at the two time points could be indicative of the same "memory update mechanism" in the sense that the retrieval-extinction procedure induces a short-term memory suppression that serves as the basis for the longer-term memory suppression (i.e., the reconsolidation effect). My pushback on this point is based on the notion of what constitutes a memory update mechanism; and is motivated by what I take to be a rather loose use of language/terminology in the reconsolidation literature and this paper specifically (for examples, see the title of the paper and line 2 of the abstract).

      As we mentioned previously, the term “mechanism” might have different connotations for different researchers. We aim to report a novel memory deficit following the retrieval-extinction paradigm, which differed significantly from the purported reconsolidation related long-term fear amnesia in terms of its timescale, cue-specificity and thought-control ability. Further TMS study confirmed that the intact dlPFC function is necessary for the short-term memory deficit. It’s based on these results we proposed that the short-term fear amnesia might be related to a different cognitive “mechanism”. As mentioned above, we now clarify what we mean by “mechanism” in the abstract and introduction (lines 31-34, 97-101).

      Reviewer #2 (Public review):

      The fear acquisition data is converted to a differential fear SCR and this is what is analysed (early vs late). However, the figure shows the raw SCR values for CS+ and CS- and therefore it is unclear whether acquisition was successful (despite there being an "early" vs "late" effect - no descriptives are provided).

      (1) There are still no descriptive statistics to substantiate learning in Experiment 1.

      We answered this question in our previous response letter. We are sorry that the definition of “early” and “late” trials was scattered in the manuscript. For example, we wrote “the late phase of acquisition (last 5 trials)” (Line 375-376) in the results section. Since there were 10 trials in total for the acquisition stage, we define the first 5 trials and the last 5 trials as “early” and “late” phases of the acquisition stage and explicitly added them into the first occasion “early” and “late” terms appeared (lines 316-318).

      In the results section, we did test whether the acquisition was successful in our previous manuscript (Line 316-325):

      “To assess fear acquisition across groups (Figure 1B and C), we conducted a mixed two-way ANOVA of group (reminder vs. no-reminder) x time (early vs. late part of the acquisition; first 5 and last 5 trials, correspondingly) on the differential fear SCR. Our results showed a significant main effect of time (early vs. late; F<sub>1,55</sub> \= 6.545, P \= 0.013, η<sup>2</sup> \= 0.106), suggesting successful fear acquisition in both groups. There was no main effect of group (reminder vs. no-reminder) or the group x time interaction (group: F<sub>1,55</sub> \= 0.057, P \= 0.813, η<sup>2</sup> \= 0.001; interaction: F<sub>1,55</sub> \= 0.066, P \= 0.798, η<sup>2</sup> \= 0.001), indicating similar levels of fear acquisition between two groups. Post-hoc t-tests confirmed that the fear responses to the CS+ were significantly higher than that of CS- during the late part of acquisition phase in both groups (reminder group: t<sub>29</sub> \= 6.642, P < 0.001; no-reminder group: t<sub>26</sub> = 8.522, P < 0.001; Figure 1C). Importantly, the levels of acquisition were equivalent in both groups (early acquisition: t<sub>55</sub> \= -0.063, P \= 0.950; late acquisition: t<sub>55</sub> \= -0.318, P \= 0.751; Figure 1C).”

      In Experiment 1 (Test results) it is unclear whether the main conclusion stems from a comparison of the test data relative to the last extinction trial ("we defined the fear recovery index as the SCR difference between the first test trial and the last extinction trial for a specific CS") or the difference relative to the CS- ("differential fear recovery index between CS+ and CS-"). It would help the reader assess the data if Fig 1e presents all the indexes (both CS+ and CS-). In addition, there is one sentence which I could not understand "there is no statistical difference between the differential fear recovery indexes between CS+ in the reminder and no reminder groups (P=0.048)". The p value suggests that there is a difference, yet it is not clear what is being compared here. Critically, any index taken as a difference relative to the CS- can indicate recovery of fear to the CS+ or absence of discrimination relative to the CS-, so ideally the authors would want to directly compare responses to the CS+ in the reminder and no-reminder groups. In the absence of such comparison, little can be concluded, in particular if SCR CS- data is different between groups. The latter issue is particularly relevant in Experiment 2, in which the CS- seems to vary between groups during the test and this can obscure the interpretation of the result.

      (2) In the revised analyses, the authors now show that CS- changes in different groups (for example, Experiment 2) so this means that there is little to conclude from the differential scores because these depend on CS-. It is unclear whether the effects arise from CS+ performance or the differential which is subject to CS- variations.

      There was a typo in the “P = 0.048” sentence and we have corrected it in our last response letter. Also in the previous response letter, we specifically addressed how the fear recovery index was defined (also in the revised manuscript).

      In most of the fear conditioning studies, CS- trials were included as the baseline control. In turn, most of the analyses conducted also involved comparisons between different groups. Directly comparing CS+ trials across groups (or conditions) is rare. In our study 2, we showed that the CS- response decreased as a function of testing delays (30min, 1hr, 6hr and 24hr). Ideally, it would be nice to show that the CS- across groups/conditions did not change. However, even in those circumstances, comparisons are still based on the differential CS response (CS+ minus CS-), that is, the difference of difference. It is also important to note that difference score is important as CS+ alone or across conditions is difficult to interpret, especially in humans, due to noise, signal fluctuations, and irrelevant stimulus features; therefore trials-wise reference is essential to assess the CS+ in the context of a reference stimulus in each trial (after all, the baselines are different). We are listing a few influential papers in the field that the CS- responses were not particularly equivalent across groups/conditions and argue that this is a routine procedure (Kindt & Soeter 2018 Figs. 2-3; Sevenster et al., 2013 Fig. 3; Liu et al., 2014 Fig. 1; Raio et al., 2017 Fig. 2).

      In experiment 1, the findings suggest that there is a benefit of retrieval followed by extinction in a short-term reinstatement test. In Experiment 2, the same effect is observed to a cue which did not undergo retrieval before extinction (CS2+), a result that is interpreted as resulting from cue-independence, rather than a failure to replicate in a within-subjects design the observations of Experiment 1 (between-subjects). Although retrieval-induced forgetting is cue-independent (the effect on items that are suppressed [Rp-] can be observed with an independent probe), it is not clear that the current findings are similar, and thus that the strong parallels made are not warranted. Here, both cues have been extinguished and therefore been equally exposed during the critical stage.

      (3) The notion that suppression is automatic is speculative at best

      We have responded the same question in our previous revision. Please note that our results from study 1 (the comparison between reminder and no-reminder groups) was not set up to test the cue-independence hypothesis for the short-term amnesia with only one CS+. Results from both study 2 (30min condition) and study 3 confirmed the cue-independence hypothesis and therefore we believe interpreting results from study 2 as “a failure to replicate in a within-subject design of the observations of Experiment 1” is not the case.

      We agree that the proposal of automatic or unconscious memory suppression is speculative and that’s why we mentioned it in the discussion. The timescale, cue-specificity and the thought-control ability dependence of the short-term fear amnesia identified in our studies was reminiscent of the memory suppression effects reported in the previous literature. However, memory suppression typically adopted a conscious “suppression” treatment (such as the think/no-think paradigm), which was absent in the current study. However, the retrieval-induced forgetting (RIF), which is also considered a memory suppression paradigm via inhibitory control, does not require conscious effort to suppress any particular thought. Based on these results and extant literature, we raised the possibility of memory suppression as a potential mechanism. We make clear in the discussion that the suppression hypothesis and connections with RIF will require further evidence (lines 615-616):

      “future research will be needed to investigate whether the short-term effect we observed is specifically related to associative memory or the spontaneous nature of suppression as in RIF (Figure 6C).”

      (4) It still struggle with the parallels between these findings and the "limbo" literature. Here you manipulated the retention interval, whereas in the cited studies the number of extinction (exposure) was varied. These are two completely different phenomena.

      We borrowed the “limbo” term to stress the transitioning from short-term to long-term memory deficits (the 6hr test group). Merlo et al. (2014) found that memory reconsolidation and extinction were dissociable processes depending on the extent of memory retrieval. They argued that there was a “limbo” transitional state, where neither the reconsolidation nor the extinction process was engaged. Our results suggest that at the test delay of 6hr, neither the short-term nor the long-term effect was present, signaling a “transitional” state after which the short-term memory deficit wanes and the long-term deficit starts to take over. We make this idea more explicit as follows (lines 622-626):

      “These works identified important “boundary conditions” of memory retrieval in affecting the retention of the maladaptive emotional memories. In our study, however, we showed that even within a boundary condition previously thought to elicit memory reconsolidation, mnemonic processes other than reconsolidation could also be at work, and these processes jointly shape the persistence of fear memory.”

      (5) My point about the data problematic for the reconsolidation (and consolidation) frameworks is that they observed memory in the absence of the brain substrates that are needed for memory to be observed. The answer did not address this. I do not understand how the latent cause model can explain this, if the only difference is the first ITI. Wouldn't participants fail to integrate extinction with acquisition with a longer ITI?

      We take the sentence “they observed memory in the absence of the brain substrates that are needed for memory to be observed” as referring to the long-term memory deficit in our study. As we responded before, the aim of this manuscript was not about investigating the brain substrates involved in memory reconsolidation (or consolidation). Using a memory retrieval-extinction paradigm, we discovered a novel short-term memory effect, which differed from the purported reconsolidation effect in terms of timescale, cue-specificity and thought-control ability dependence. We further showed that both memory retrieval and intact dlPFC functions were necessary to observe the short-term memory deficit effect. Therefore, we conclude that the brain mechanism involved in such an effect should be different from the one related to the purported reconsolidation effect. We make this idea more explicit as follows (lines 546-547):

      “Therefore, findings of the short-term fear amnesia suggest that the reconsolidation framework falls short to accommodate this more immediate effect (Figure 6A and B).”

      Whilst I could access the data in the OFS site, I could not make sense of the Matlab files as there is no signposting indicating what data is being shown in the files. Thus, as it stands, there is no way of independently replicating the analyses reported.

      (6) The materials in the OSF site are the same as before, they haven't been updated.

      Last time we thought the main issue was the OSF site not being publicly accessible and thus made it open to all visitors. We have added descriptive file to explain the variables to help visitors to replicate the analyses we took.

      (7) Concerning supplementary materials, the robustness tests are intended to prove that you 1) can get the same results by varying the statistical models or 2) you can get the same results when you include all participants. Here authors have done both so this does not help. Also, in the rebuttal letter, they stated "Please note we did not include non-learners in these analyses " which contradicts what is stated in the figure captions "(learners + non learners)"

      In the supplementary materials, we did the analyses of varying the statistical models and including both learners and non-learners separately, instead of both. In fact, in the supplementary material Figs. 1 & 2, we included all the participants and performed similar analysis as in the main text and found similar results (learners + non-learners). Also, in the text of the supplementary material, we used a different statistical analysis method to only learners (analyzing subjects reported in the main text using a different method) and achieved similar results. We believe this is exactly what the reviewer suggested us to do. Also there seems to be a misunderstanding for the "Please note we did not include non-learners in these analyses" sentence in the rebuttal letter. As the reviewer can see, the full sentence read “Please note we did not include non-learners in these analyses (the texts of the supplementary materials)”. We meant to express that the Figures and texts in the supplementary material reflect two approaches: 1) Figures depicting re-analysis with all the included subjects (learners + non learners); 2) Text describing different analysis with learners. We added clarifications to emphasize these approaches in the supplementary materials.

      (8) Finally, the literature suggesting that reconsolidation interference "eliminates" a memory is not substantiated by data nor in line with current theorising, so I invite a revision of these strong claims.

      We agree and have toned down the strong claims.

      Overall, I conclude that the revised manuscript did not address my main concerns.

      In both rounds of responses, we tried our best to address the reviewer’s concerns. We hope that the clarifications in this letter and revisions in the text address the remaining concerns. Thank you for your feedback.

      Reference:

      Kindt, M. and Soeter, M. 2018. Pharmacologically induced amnesia for learned fear is time and sleep dependent. Nat Commun, 9, 1316.

      Liu, J., Zhao, L., Xue, Y., Shi, J., Suo, L., Luo, Y., Chai, B., Yang, C., Fang, Q., Zhang, Y., Bao, Y., Pickens, C. L. and Lu, L. 2014. An unconditioned stimulus retrieval extinction procedure to prevent the return of fear memory. Biol Psychiatry, 76, 895-901.

      Raio, C. M., Hartley, C. A., Orederu, T. A., Li, J. and Phelps, E. A. 2017. Stress attenuates the flexible updating of aversive value. Proc Natl Acad Sci U S A, 114, 11241-11246.

      Sevenster, D., Beckers, T., & Kindt, M. 2013. Prediction error governs pharmacologically induced amnesia for learned fear. Science (New York, N.Y.), 339(6121), 830–833.

    1. He said that the media themselves were far more important than any content they carried.

      The argument that form is more important than the message is interesting. Considering that the type of media impacts how an idea is presented and controls the way the viewer/reader/audience can engage with the content, it makes sense McLuhan would believe this claim.

    1. Political theorist Langdon Winner once asked whether artifacts can have politics. They can and AI systems are no exception. They encode assumptions about what counts as intelligence and whose labor counts as valuable. The more we rely on algorithms, the more we normalize their values: automation, prediction, standardization, and corporate dependency. Eventually these priorities fade from view and come to seem natural—“just the way things are.”

      Infrastructural. Mundane.

    1. Generative AI can mimic some of the work done by historians and history educators. This should not be mistaken for teaching or for learning.

      AI can write paragraphs that look like real history, but that doesn’t mean students are actually learning anything. Real learning happens when a person thinks, analyzes evidence, and makes sense of the past. Something AI cannot do.

    1. eLife Assessment

      This study offers valuable insights into how humans detect and adapt to regime shifts, highlighting distinct contributions of the frontoparietal network and ventromedial prefrontal cortex to sensitivity to signal diagnosticity and transition probabilities. The combination of an innovative task design, behavioral modeling, and model-based fMRI analyses provides a solid foundation for the conclusions; however, the neuroimaging results have several limitations, particularly a potential confound between the posterior probability of a switch and the passage of time that may not be fully controlled by including trial number as a regressor. The control experiments intended to address this issue also appear conceptually inconsistent and, at the behavioral level, while informing participants of conditional probabilities rather than requiring learning is theoretically elegant, such information is difficult to apply accurately, as shown by well-documented challenges with conditional reasoning and base-rate neglect. Expressing these probabilities as natural frequencies rather than percentages may have improved comprehension. Overall, the study advances understanding of belief updating under uncertainty but would benefit from more intuitive probabilistic framing and stronger control of temporal confounds in future work.

    2. Reviewer #1 (Public review):

      Summary:

      The study examines human biases in a regime-change task, in which participants have to report the probability of a regime change in the face of noisy data. The behavioral results indicate that humans display systematic biases, in particular, overreaction in stable but noisy environments and underreaction in volatile settings with more certain signals. fMRI results suggest that a frontoparietal brain network is selectively involved in representing subjective sensitivity to noise, while the vmPFC selectively represents sensitivity to the rate of change.

      Strengths:

      - The study relies on a task that measures regime-change detection primarily based on descriptive information about the noisiness and rate of change. This distinguishes the study from prior work using reversal-learning or change-point tasks in which participants are required to learn these parameters from experiences. The authors discuss these differences comprehensively.

      - The study uses a simple Bayes-optimal model combined with model fitting, which seems to describe the data well. The model is comprehensively validated.

      - The authors apply model-based fMRI analyses that provide a close link to behavioral results, offering an elegant way to examine individual biases.

      Weaknesses:

      The authors have adequately addressed most of my prior concerns.

      My only remaining comment concerns the z-test of the correlations. I agree with the non-parametric test based on bootstrapping at the subject level, providing evidence for significant differences in correlations within the left IFG and IPS.

      However, the parametric test seems inadequate to me. The equation presented is described as the Fisher z-test, but the numerator uses the raw correlation coefficients (r) rather than the Fisher-transformed values (z). To my understanding, the subtraction should involve the Fisher z-scores, not the raw correlations.

      More importantly, the Fisher z-test in its standard form assumes that the correlations come from independent samples, as reflected in the denominator (which uses the n of each independent sample). However, in my opinion, the two correlations are not independent but computed within-subject. In such cases, parametric tests should take into account the dependency. I believe one appropriate method for the current case (correlated correlation coefficients sharing a variable [behavioral slope]) is explained here:

      Meng, X.-l., Rosenthal, R., & Rubin, D. B. (1992). Comparing correlated correlation coefficients. Psychological Bulletin, 111(1), 172-175. https://doi.org/10.1037/0033-2909.111.1.172

      It should be implemented here:

      Diedenhofen B, Musch J (2015) cocor: A Comprehensive Solution for the Statistical Comparison of Correlations. PLoS ONE 10(4): e0121945. https://doi.org/10.1371/journal.pone.0121945

      My recommendation is to verify whether my assumptions hold, and if so, perform a test that takes correlated correlations into account. Or, to focus exclusively on the non-parametric test.

      In any case, I recommend a short discussion of these findings and how the authors interpret that some of the differences in correlations are not significant.

    3. Reviewer #3 (Public review):

      This study concerns how observers (human participants) detect changes in the statistics of their environment, termed regime shifts. To make this concrete, a series of 10 balls are drawn from an urn that contains mainly red or mainly blue balls. If there is a regime shift, the urn is changed over (from mainly red to mainly blue) at some point in the 10 trials. Participants report their belief that there has been a regime shift as a % probability. Their judgement should (mathematically) depend on the prior probability of a regime shift (which is set at one of three levels) and the strength of evidence (also one of three levels, operationalized as the proportion of red balls in the mostly-blue urn and vice versa). Participants are directly instructed of the prior probability of regime shift and proportion of red balls, which are presented on-screen as numerical probabilities. The task therefore differs from most previous work on this question in that probabilities are instructed rather than learned by observation, and beliefs are reported as numerical probabilities rather than being inferred from participants' choice behaviour (as in many bandit tasks, such as Behrens 2007 Nature Neurosci).

      The key behavioural finding is that participants over-estimate the prior probability of regime change when it is low, and under estimate it when it is high; and participants over-estimate the strength of evidence when it is low and under-estimate it when it is high. In other words participants make much less distinction between the different generative environments than an optimal observer would. This is termed 'system neglect'. A neuroeconomic-style mathematical model is presented and fit to data.

      Functional MRI results how that strength of evidence for a regime shift (roughly, the surprise associated with a blue ball from an apparently red urn) is associated with activity in the frontal-parietal orienting network. Meanwhile, at time-points where the probability of a regime shift is high, there is activity in another network including vmPFC. Both networks show individual differences effects, such that people who were more sensitive to strength of evidence and prior probability show more activity in the frontal-parietal and vmPFC-linked networks respectively.

      Strengths

      (1) The study provides a different task for looking at change-detection and how this depends on estimates of environmental volatility and sensory evidence strength, in which participants are directly and precisely informed of the environmental volatility and sensory evidence strength rather than inferring them through observation as in most previous studies<br /> (2) Participants directly provide belief estimates as probabilities rather than experimenters inferring them from choice behaviour as in most previous studies<br /> (3) The results are consistent with well-established findings that surprising sensory events activate the frontal-parietal orienting network whilst updating of beliefs about the word ('regime shift') activates vmPFC.

      Weaknesses

      (1) The use of numerical probabilities (both to describe the environments to participants, and for participants to report their beliefs) may be problematic because people are notoriously bad at interpreting probabilities presented in this way, and show poor ability to reason with this information (see Kahneman's classic work on probabilistic reasoning, and how it can be improved by using natural frequencies). Therefore the fact that, in the present study, people do not fully use this information, or use it inaccurately, may reflect the mode of information delivery.

      (2) Although a very precise model of 'system neglect' is presented, many other models could fit the data.

      For example, you would get similar effects due to attraction of parameter estimates towards a global mean - essentially application of a hyper-prior in which the parameters applied by each participant in each block are attracted towards the experiment-wise mean values of these parameters. For example, the prior probability of regime shift ground-truth values [0.01, 0.05, 0.10] are mapped to subjective values of [0.037, 0.052, 0.069]; this would occur if observers apply a hyper-prior that the probability of regime shift is about 0.05 (the average value over all blocks). This 'attraction to the mean' is a well-established phenomenon and cannot be ruled out with the current data (I suppose you could rule it out by comparing to another dataset in which the mean ground-truth value was different).

      More generally, any model in which participants don't fully use the numerical information they were given would produce apparent 'system neglect'. Four qualitatively different example reasons are: 1. Some individual participants completely ignored the probability values given. 2. Participants did not ignore the probability values given, but combined them with a hyperprior as above. 3. Participants had a reporting bias where their reported beliefs that a regime-change had occurred tend to be shifted towards 50% (rather than reporting 'confident' values such 5% or 95%). 4. Participants underweighted probability outliers resulting in underweighting of evidence in the 'high signal diagnosticity' environment (10.1016/j.neuron.2014.01.020 )

      In summary I agree that any model that fits the data would have to capture the idea that participants don't differentiate between the different environments as much as they should, but I think there are a number of qualitatively different reasons why they might do this - of which the above are only examples - hence I find it problematic that the authors present the behaviour as evidence for one extremely specific model.

      (3) Despite efforts to control confounds in the fMRI study, including two control experiments, I think some confounds remain.

      For example, a network of regions is presented as correlating with the cumulative probability that there has been a regime shift in this block of 10 samples (Pt). However, regardless of the exact samples shown, doesn't Pt always increase with sample number (as by the time of later samples, there have been more opportunities for a regime shift)? Unless this is completely linear, the effect won't be controlled by including trial number as a co-regressor (which was done).

      On the other hand, two additional fMRI experiments are done as control experiments and the effect of Pt in the main study is compared to Pt in these control experiments. Whilst I admire the effort in carrying out control studies, I can't understand how these particular experiment are useful controls. For example in experiment 3 participants simply type in numbers presented on the screen - how can we even have an estimate of Pt from this task?

      (4) The Discussion is very long, and whilst a lot of related literature is cited, I found it hard to pin down within the discussion, what the key contributions of this study are. In my opinion it would be better to have a short but incisive discussion highlighting the advances in understanding that arise from the current study, rather than reviewing the field so broadly.

      Editors’ note: Reviewer #2 was unavailable to re-review the manuscript. Reviewer #3 was added for this round of review to ensure two reviewers and because of their expertise in the computational and modelling aspects of the work.

    4. Author response:

      The following is the authors’ response to the current reviews.

      eLife Assessment<br /> This study offers valuable insights into how humans detect and adapt to regime shifts, highlighting distinct contributions of the frontoparietal network and ventromedial prefrontal cortex to sensitivity to signal diagnosticity and transition probabilities. The combination of an innovative task design, behavioral modeling, and model-based fMRI analyses provides a solid foundation for the conclusions; however, the neuroimaging results have several limitations, particularly a potential confound between the posterior probability of a switch and the passage of time that may not be fully controlled by including trial number as a regressor. The control experiments intended to address this issue also appear conceptually inconsistent and, at the behavioral level, while informing participants of conditional probabilities rather than requiring learning is theoretically elegant, such information is difficult to apply accurately, as shown by well-documented challenges with conditional reasoning and base-rate neglect. Expressing these probabilities as natural frequencies rather than percentages may have improved comprehension. Overall, the study advances understanding of belief updating under uncertainty but would benefit from more intuitive probabilistic framing and stronger control of temporal confounds in future work.

      We thank the editors for the assessment. The editor added several limitations based on the new reviewer 3 in this round, which we address below.

      With regard to temporal confounds, we clarified in the main text and response to Reviewer 3 that we had already addressed the potential confound between posterior probability of a switch and passage of time in GLM-2 with the inclusion of intertemporal prior. After adding intertemporal prior in the GLM, we still observed the same fMRI results on probability estimates. In addition, we did two other robustness checks, which we mentioned in the manuscript.

      With regard to response mode (probability estimation rather than choice or indicating natural frequencies), we wish to point out that the in previous research by Massey and Wu (2005), which the current study was based on, the concern of participants showing system-neglect tendencies due to the mode of information delivery, namely indicating beliefs through reporting probability estimates rather than through choice or other response mode was addressed. Massy and Wu (2005, Study 3) found the same biases when participants performed a choice task that did not require them to indicate probability estimates.

      With regard to the control experiments, the control experiments in fact were not intended to address the confounds between posterior probability and passage of time. Rather, they aimed to address whether the neural findings were unique to change detection (Experiment 2) and to address visual and motor confounds (Experiment 3). These and the results of the control experiments were mentioned on page 18-19.

      Finally, we wish to highlight that we had performed detailed model comparisons after reviewer 2’s suggestions. Although reviewer 2 was unable to re-review the manuscript, we believe this provides insight into the literature on change detection. See “Incorporating signal dependency into system-neglect model led to better models for regime-shift detection” (p.27-30). The model comparison showed that system-neglect models that incorporate signal dependency are better models than the original system-neglect model in describing participants probability estimates. This suggests that people respond to change-consistent and change-inconsistent signals differently when judging whether the regime had changed. This was not reported in previous behavioral studies and was largely inspired by the neural finding on signal dependency in the frontoparietal cortex. It indicates that neural findings can provide novel insights into computational modeling of behavior.           

      Public Reviews:

      Reviewer #1 (Public review):

      Summary:

      The study examines human biases in a regime-change task, in which participants have to report the probability of a regime change in the face of noisy data. The behavioral results indicate that humans display systematic biases, in particular, overreaction in stable but noisy environments and underreaction in volatile settings with more certain signals. fMRI results suggest that a frontoparietal brain network is selectively involved in representing subjective sensitivity to noise, while the vmPFC selectively represents sensitivity to the rate of change.

      Strengths:

      - The study relies on a task that measures regime-change detection primarily based on descriptive information about the noisiness and rate of change. This distinguishes the study from prior work using reversal-learning or change-point tasks in which participants are required to learn these parameters from experiences. The authors discuss these differences comprehensively.

      - The study uses a simple Bayes-optimal model combined with model fitting, which seems to describe the data well. The model is comprehensively validated.

      - The authors apply model-based fMRI analyses that provide a close link to behavioral results, offering an elegant way to examine individual biases.

      We thank the reviewer for the comments.

      Weaknesses:

      The authors have adequately addressed most of my prior concerns.

      We thank the reviewer for recognizing our effort in addressing your concerns.

      My only remaining comment concerns the z-test of the correlations. I agree with the non-parametric test based on bootstrapping at the subject level, providing evidence for significant differences in correlations within the left IFG and IPS.

      However, the parametric test seems inadequate to me. The equation presented is described as the Fisher z-test, but the numerator uses the raw correlation coefficients (r) rather than the Fisher-transformed values (z). To my understanding, the subtraction should involve the Fisher z-scores, not the raw correlations.

      More importantly, the Fisher z-test in its standard form assumes that the correlations come from independent samples, as reflected in the denominator (which uses the n of each independent sample). However, in my opinion, the two correlations are not independent but computed within-subject. In such cases, parametric tests should take into account the dependency. I believe one appropriate method for the current case (correlated correlation coefficients sharing a variable [behavioral slope]) is explained here:

      Meng, X.-l., Rosenthal, R., & Rubin, D. B. (1992). Comparing correlated correlation coefficients. Psychological Bulletin, 111(1), 172-175. https://doi.org/10.1037/0033-2909.111.1.172

      It should be implemented here:

      Diedenhofen B, Musch J (2015) cocor: A Comprehensive Solution for the Statistical Comparison of Correlations. PLoS ONE 10(4): e0121945. https://doi.org/10.1371/journal.pone.0121945

      My recommendation is to verify whether my assumptions hold, and if so, perform a test that takes correlated correlations into account. Or, to focus exclusively on the non-parametric test.

      In any case, I recommend a short discussion of these findings and how the authors interpret that some of the differences in correlations are not significant.

      Thank you for the careful check. Yes. This was indeed a mistake from us. We also agree that the two correlations are not independent. Therefore, we modified the test that accounts for dependent correlations by following Meng et al. (1992) suggested by the reviewer.

      We referred to the correlation between neural and behavioral sensitivity at change-consistent (blue) signals as , and that at change-inconsistent (red) signals as 𝑟<sub>𝑟𝑒𝑑</sub>. To statistically compare these two correlations, we adopted the approach of Meng et al. (1992), which specifically tests differences between dependent correlations according to the following equation

      where  is the number of subjects, 𝑧<sub>𝑟𝑖</sub> is the Fisher z-transformed value of 𝑟<sub>𝑖</sub>, 𝑟<sub>1</sub> = 𝑟<sub>𝑏𝑙𝑢𝑒</sub> and 𝑟<sub>2</sub> = 𝑟<sub>𝑟𝑒𝑑</sub>. 𝑟<sub>𝑥</sub> is the correlation between the neural sensitivity at change-consistent signals and change-inconsistent signals.

      Where is the mean of the , and 𝑓 should be set to 1 if > 1.

      We found that among the five ROIs in the frontoparietal network, two of them, namely the left IFG and left IPS, the difference in correlation was significant (one-tailed z test; left IFG: 𝑧 = 1.8908, 𝑝 = 0.0293; left IPS: 𝑧 = 2.2584, 𝑝 = 0.0049). For the remaining three ROIs, the difference in correlation was not significant (dmPFC: 𝑧 = 0.9522, 𝑝 = 0.1705; right IFG: 𝑧 = 0.9860, 𝑝 = 0.1621; right IPS: 𝑧 = 1.4833, 𝑝 = 0.0690). We chose one-tailed test because we already know the correlation under the blue signals was significantly greater than 0. These updated results are consistent with the nonparametric tests we had already performed and we will update them in the revised manuscript.

      Reviewer #3 (Public review):

      This study concerns how observers (human participants) detect changes in the statistics of their environment, termed regime shifts. To make this concrete, a series of 10 balls are drawn from an urn that contains mainly red or mainly blue balls. If there is a regime shift, the urn is changed over (from mainly red to mainly blue) at some point in the 10 trials. Participants report their belief that there has been a regime shift as a % probability. Their judgement should (mathematically) depend on the prior probability of a regime shift (which is set at one of three levels) and the strength of evidence (also one of three levels, operationalized as the proportion of red balls in the mostly-blue urn and vice versa). Participants are directly instructed of the prior probability of regime shift and proportion of red balls, which are presented on-screen as numerical probabilities. The task therefore differs from most previous work on this question in that probabilities are instructed rather than learned by observation, and beliefs are reported as numerical probabilities rather than being inferred from participants' choice behaviour (as in many bandit tasks, such as Behrens 2007 Nature Neurosci).

      The key behavioural finding is that participants over-estimate the prior probability of regime change when it is low, and under estimate it when it is high; and participants over-estimate the strength of evidence when it is low and under-estimate it when it is high. In other words participants make much less distinction between the different generative environments than an optimal observer would. This is termed 'system neglect'. A neuroeconomic-style mathematical model is presented and fit to data.

      Functional MRI results how that strength of evidence for a regime shift (roughly, the surprise associated with a blue ball from an apparently red urn) is associated with activity in the frontal-parietal orienting network. Meanwhile, at time-points where the probability of a regime shift is high, there is activity in another network including vmPFC. Both networks show individual differences effects, such that people who were more sensitive to strength of evidence and prior probability show more activity in the frontal-parietal and vmPFC-linked networks respectively.

      We thank the reviewer for the overall descriptions of the manuscript.

      Strengths:

      (1) The study provides a different task for looking at change-detection and how this depends on estimates of environmental volatility and sensory evidence strength, in which participants are directly and precisely informed of the environmental volatility and sensory evidence strength rather than inferring them through observation as in most previous studies

      (2) Participants directly provide belief estimates as probabilities rather than experimenters inferring them from choice behaviour as in most previous studies<br /> (3) The results are consistent with well-established findings that surprising sensory events activate the frontal-parietal orienting network whilst updating of beliefs about the word ('regime shift') activates vmPFC.

      Thank you for these assessments.

      Weaknesses:

      (1) The use of numerical probabilities (both to describe the environments to participants, and for participants to report their beliefs) may be problematic because people are notoriously bad at interpreting probabilities presented in this way, and show poor ability to reason with this information (see Kahneman's classic work on probabilistic reasoning, and how it can be improved by using natural frequencies). Therefore the fact that, in the present study, people do not fully use this information, or use it inaccurately, may reflect the mode of information delivery.

      We appreciate the reviewer’s concern on this issue. The concern was addressed in Massey and Wu (2005) as participants performed a choice task in which they were not asked to provide probability estimates (Study 3 in Massy and Wu, 2005). Instead, participants in Study 3 were asked to predict the color of the ball before seeing a signal. This was a more intuitive way of indicating his or her belief about regime shift. The results from the choice task were identical to those found in the probability estimation task (Study 1 in Massey and Wu). We take this as evidence that the system-neglect behavior the participants showed was less likely to be due to the mode of information delivery.

      (2) Although a very precise model of 'system neglect' is presented, many other models could fit the data.

      For example, you would get similar effects due to attraction of parameter estimates towards a global mean - essentially application of a hyper-prior in which the parameters applied by each participant in each block are attracted towards the experiment-wise mean values of these parameters. For example, the prior probability of regime shift ground-truth values [0.01, 0.05, 0.10] are mapped to subjective values of [0.037, 0.052, 0.069]; this would occur if observers apply a hyper-prior that the probability of regime shift is about 0.05 (the average value over all blocks). This 'attraction to the mean' is a well-established phenomenon and cannot be ruled out with the current data (I suppose you could rule it out by comparing to another dataset in which the mean ground-truth value was different).

      We thank the reviewer for this comment. It is true that the system-neglect model is not entirely inconsistent with regression to the mean, regardless of whether the implementation has a hyper prior or not. In fact, our behavioral measure of sensitivity to transition probability and signal diagnosticity, which we termed the behavioral slope, is based on linear regression analysis. In general, the modeling approach in this paper is to start from a generative model that defines ideal performance and consider modifying the generative model when systematic deviations in actual performance from the ideal is observed. In this approach, a generative model with hyper-prior would be more complex to begin with, and a regression to the mean idea by itself does not generate a priori predictions.

      More generally, any model in which participants don't fully use the numerical information they were given would produce apparent 'system neglect'. Four qualitatively different example reasons are: 1. Some individual participants completely ignored the probability values given. 2. Participants did not ignore the probability values given, but combined them with a hyperprior as above. 3. Participants had a reporting bias where their reported beliefs that a regime-change had occurred tend to be shifted towards 50% (rather than reporting 'confident' values such 5% or 95%). 4. Participants underweighted probability outliers resulting in underweighting of evidence in the 'high signal diagnosticity' environment (10.1016/j.neuron.2014.01.020 )

      In summary I agree that any model that fits the data would have to capture the idea that participants don't differentiate between the different environments as much as they should, but I think there are a number of qualitatively different reasons why they might do this - of which the above are only examples - hence I find it problematic that the authors present the behaviour as evidence for one extremely specific model.

      Thank you for raising this point. The modeling principle we adopt is the following. We start from the normative model—the Bayesian model—that defined what normative behavior should look like. We compared participants’ behavior with the Bayesian model and found systematic deviations from it. To explain those systematic deviations, we considered modeling options within the confines of the same modeling framework. In other words, we considered a parameterized version of the Bayesian model, which is the system-neglect model and examined through model comparison the best modeling choice. This modeling approach is not uncommon, and many would agree this is the standard approach in economics and psychology. For example, Kahneman and Tversky adopted this approach when proposing prospect theory, a modification of expected utility theory where expected utility theory can be seen as one specific model for how utility of an option should be computed.

      (3) Despite efforts to control confounds in the fMRI study, including two control experiments, I think some confounds remain.

      For example, a network of regions is presented as correlating with the cumulative probability that there has been a regime shift in this block of 10 samples (Pt). However, regardless of the exact samples shown, doesn't Pt always increase with sample number (as by the time of later samples, there have been more opportunities for a regime shift)? Unless this is completely linear, the effect won't be controlled by including trial number as a co-regressor (which was done).

      Thank you for raising this concern. Yes, Pt always increases with sample number regardless of evidence (seeing change-consistent or change-inconsistent signals). This is captured by the ‘intertemporal prior’ in the Bayesian model, which we included as a regressor in our GLM analysis (GLM-2), in addition to Pt. In short, GLM-1 had Pt and sample number. GLM-2 had Pt, intertemporal prior, and sample number, among other regressors. And we found that, in both GLM-1 and GLM-2, both vmPFC and ventral striatum correlated with Pt.

      To make this clearer, we updated the main text to further clarify this on p.18:

      On the other hand, two additional fMRI experiments are done as control experiments and the effect of Pt in the main study is compared to Pt in these control experiments. Whilst I admire the effort in carrying out control studies, I can't understand how these particular experiment are useful controls. For example in experiment 3 participants simply type in numbers presented on the screen - how can we even have an estimate of Pt from this task?

      We thank the reviewer for this comment. The purpose of Experiment 3 was to control for visual and motor confounds. In other words, if subjects saw the similar visual layout and were just instructed to press numbers, would we observe the vmPFC, ventral striatum, and the frontoparietal network like what we did in the main experiment (Experiment 1)?

      The purpose of Experiment 2 was to establish whether what we found about Pt was unique to change detection. In Experiment 2, subjects estimated the probability that the current regime is the blue regime (just as they did in Experiment 1) except that there were no regime shifts involved. In other words, it is possible that the regions we identified were generally associated with probability estimation and not particularly about change detection. And we used Experiment 2 to examine whether this were true.

      (4) The Discussion is very long, and whilst a lot of related literature is cited, I found it hard to pin down within the discussion, what the key contributions of this study are. In my opinion it would be better to have a short but incisive discussion highlighting the advances in understanding that arise from the current study, rather than reviewing the field so broadly.

      Thank you. We received different feedbacks from previous reviews on what to include in Discussion. To address the reviewer’s concern, we will revise the Discussion to better highlight the key contributions of the current study at the beginning of Discussion.

      Recommendations for the authors:

      Reviewer #3 (Recommendations for the authors):

      Many of the figures are too tiny - the writing is very small, as are the pictures of brains. I'd suggest adjusting these so they will be readable without enlarging.

      Thank you. We will enlarge the figures to make them more readable.


      The following is the authors’ response to the original reviews.

      Public Reviews:

      Reviewer #1 (Public review):

      Summary:

      The study examines human biases in a regime-change task, in which participants have to report the probability of a regime change in the face of noisy data. The behavioral results indicate that humans display systematic biases, in particular, overreaction in stable but noisy environments and underreaction in volatile settings with more certain signals. fMRI results suggest that a frontoparietal brain network is selectively involved in representing subjective sensitivity to noise, while the vmPFC selectively represents sensitivity to the rate of change.

      Strengths:

      (1) The study relies on a task that measures regime-change detection primarily based on descriptive information about the noisiness and rate of change. This distinguishes the study from prior work using reversal-learning or change-point tasks in which participants are required to learn these parameters from experiences. The authors discuss these differences comprehensively.

      Thank you for recognizing our contribution to the regime-change detection literature and our effort in discussing our findings in relation to the experience-based paradigms.

      (2) The study uses a simple Bayes-optimal model combined with model fitting, which seems to describe the data well.

      Thank you for recognizing the contribution of our Bayesian framework and systemneglect model.

      (3) The authors apply model-based fMRI analyses that provide a close link to behavioral results, offering an elegant way to examine individual biases.

      Thank you for recognizing our execution of model-based fMRI analyses and effort in using those analyses to link with behavioral biases.

      Weaknesses:

      My major concern is about the correlational analysis in the section "Under- and overreactions are associated with selectivity and sensitivity of neural responses to system parameters", shown in Figures 5c and d (and similarly in Figure 6). The authors argue that a frontoparietal network selectively represents sensitivity to signal diagnosticity, while the vmPFC selectively represents transition probabilities. This claim is based on separate correlational analyses for red and blue across different brain areas. The authors interpret the finding of a significant correlation in one case (blue) and an insignificant correlation (red) as evidence of a difference in correlations (between blue and red) but don't test this directly. This has been referred to as the "interaction fallacy" (Niewenhuis et al., 2011; Makin & Orban de Xivry 2019). Not directly testing the difference in correlations (but only the differences to zero for each case) can lead to wrong conclusions. For example, in Figure 5c, the correlation for red is r = 0.32 (not significantly different from zero) and r = 0.48 (different from zero). However, the difference between the two is 0.1, and it is likely that this difference itself is not significant. From a statistical perspective, this corresponds to an interaction effect that has to be tested directly. It is my understanding that analyses in Figure 6 follow the same approach.

      Relevant literature on this point is:

      Nieuwenhuis, S, Forstmann, B & Wagenmakers, EJ (2011). Erroneous analyses of interactions in neuroscience: a problem of significance. Nat Neurosci 14, 11051107. https://doi.org/10.1038/nn.2886

      Makin TR, Orban de Xivry, JJ (2019). Science Forum: Ten common statistical mistakes to watch out for when writing or reviewing a manuscript. eLife 8:e48175. https://doi.org/10.7554/eLife.48175

      There is also a blog post on simulation-based comparisons, which the authors could check out: https://garstats.wordpress.com/2017/03/01/comp2dcorr/

      I recommend that the authors carefully consider what approach works best for their purposes. It is sometimes recommended to directly compare correlations based on Monte-Carlo simulations (cf Makin & Orban). It might also be appropriate to run a regression with the dependent variable brain activity (Y) and predictors brain area (X) and the model-based term of interest (Z). In this case, they could include an interaction term in the model:

      Y = \beta_0 + \beta_1 \cdot X + \beta_2 \cdot Z + \beta_3 \cdot X \cdot Z

      The interaction term reflects if the relationship between the model term Z and brain activity Y is conditional on the brain area of interest X.

      Thank you for the suggestion. In response, we tested for the difference in correlation both parametrically and nonparametrically. The results were identical. In the parametric test, we used the Fisher z transformation to transform the difference in correlation coefficients to the z statistic. That is, for two correlation coefficients, 𝑟<sub>1</sub> (with sample size 𝑛<sub>1</sub>) and 𝑟<sub>2</sub>, (with sample size 𝑛<sub>2</sub>), the z statistic of the difference in correlation is given by

      We referred to the correlation between neural and behavioral sensitivity at change-consistent (blue) signals as 𝑟<sub>𝑏𝑙𝑢𝑒</sub>, and that at change-inconsistent (red) signals as 𝑟<sub>𝑟𝑒𝑑</sub>. For the Fisher z transformation 𝑟<sub>1</sub>= 𝑟<sub>𝑏𝑙𝑢𝑒</sub> and 𝑟<sub>2</sub> \= 𝑟<sub>𝑟𝑒𝑑</sub>. We found that among the five ROIs in the frontoparietal network, two of them, namely the left IFG and left IPS, the difference in correlation was significant (one-tailed z test; left IFG: 𝑧 = 1.8355, 𝑝 =0.0332; left IPS: 𝑧 = 2.3782, 𝑝 = 0.0087). For the remaining three ROIs, the difference in correlation was not significant (dmPFC: 𝑧 = 0.7594, 𝑝 = 0.2238; right IFG: 𝑧 = 0.9068, 𝑝 = 0.1822; right IPS: 𝑧 = 1.3764, 𝑝 = 0.0843). We chose one-tailed test because we already know the correlation under the blue signals was significantly greater than 0.

      In the nonparametric test, we performed nonparametric bootstrapping to test for the difference in correlation (Efron & Tibshirani, 1994). We resampled with replacement the dataset (subject-wise) and used the resampled dataset to compute the difference in correlation. We then repeated the above for 100,000 times so as to estimate the distribution of the difference in correlation coefficients, tested for significance and estimated p-value based on this distribution. Consistent with our parametric tests, here we also found that the difference in correlation was significant in left IFG and left IPS (left IFG: 𝑟<sub>𝑏𝑙𝑢𝑒</sub> − 𝑟<sub>𝑟𝑒𝑑</sub> \= 0.46, 𝑝 = 0.0496; left IPS: 𝑟<sub>𝑏𝑙𝑢𝑒</sub> − 𝑟<sub>𝑟𝑒𝑑</sub> \= 0.5306, 𝑝 = 0.0041), but was not significant in dmPFC, right IFG, and right IPS (dmPFC: 𝑟<sub>𝑏𝑙𝑢𝑒</sub> − 𝑟<sub>𝑟𝑒𝑑</sub> \= 0.1634, 𝑝 = 0.1919; right IFG: 𝑟<sub>𝑏𝑙𝑢𝑒</sub> − 𝑟<sub>𝑟𝑒𝑑</sub> \= 0.2123, 𝑝 = 0.1681; right IPS: 𝑟<sub>𝑏𝑙𝑢𝑒</sub> − 𝑟<sub>𝑟𝑒𝑑</sub> \= 0.3434, 𝑝 = 0.0631).

      In summary, we found that neural sensitivity to signal diagnosticity in the frontoparietal network measured at change-consistent signals significantly correlated with individual subjects’ behavioral sensitivity to signal diagnosticity (𝑟<sub>𝑏𝑙𝑢𝑒</sub>). By contrast, neural sensitivity to signal diagnosticity measured at change-inconsistent did not significantly correlate with behavioral sensitivity (𝑟<sub>𝑟𝑒𝑑</sub>). The difference in correlation, 𝑟<sub>𝑏𝑙𝑢𝑒</sub> − 𝑟<sub>𝑟𝑒𝑑</sub>, however, was statistically significant in some (left IPS and left IFG) but not all brain regions within the frontoparietal network.

      To incorporate these updates, we added descriptions of the methods and results in the revised manuscript. In the Results section (p.26-27):

      “We further tested, for each brain region, whether the difference in correlation was significant using both parametric and nonparametric tests (see Parametric and nonparametric tests for difference in correlation coefficients in Methods). The results were identical. In the parametric test, we used the Fisher 𝑧 transformation to transform the difference in correlation coefficients to the 𝑧 statistic. We found that among the five ROIs in the frontoparietal network, two of them, namely the left IFG and left IPS, the difference in correlation was significant (one-tailed z test; left IFG: 𝑧 = 1.8355, 𝑝 = 0.0332; left IPS: 𝑧 = 2.3782, 𝑝 = 0.0087). For the remaining three ROIs, the difference in correlation was not significant (dmPFC: 𝑧 = 0.7594, 𝑝 = 0.2238; right IFG: 𝑧 = 0.9068, 𝑝 = 0.1822; right IPS: 𝑧 = 1.3764, 𝑝 = 0.0843). We chose one-tailed test because we already know the correlation under change-consistent signals was significantly greater than 0. In the nonparametric test, we performed nonparametric bootstrapping to test for the difference in correlation. We referred to the correlation between neural and behavioral sensitivity at change-consistent (blue) signals as 𝑟<sub>𝑏𝑙𝑢𝑒</sub>, and that at change-inconsistent (red) signals as 𝑟<sub>𝑟𝑒𝑑</sub>. Consistent with the parametric tests, we also found that the difference in correlation was significant in left IFG and left IPS (left IFG: 𝑟<sub>𝑏𝑙𝑢𝑒</sub> − 𝑟<sub>𝑟𝑒𝑑</sub> \= 0.46, 𝑝 = 0.0496; left IPS: 𝑟<sub>𝑏𝑙𝑢𝑒</sub> − 𝑟<sub>𝑟𝑒𝑑</sub> \= 0.5306, 𝑝 = 0.0041), but was not significant in dmPFC, right IFG, and right IPS (dmPFC: 𝑟<sub>𝑏𝑙𝑢𝑒</sub> − 𝑟<sub>𝑟𝑒𝑑</sub> \=0.1634, 𝑝 = 0.1919; right IFG: 𝑟<sub>𝑏𝑙𝑢𝑒</sub> − 𝑟<sub>𝑟𝑒𝑑</sub> \= 0.2123, 𝑝 = 0.1681; right IPS: 𝑟<sub>𝑏𝑙𝑢𝑒</sub> − 𝑟<sub>𝑟𝑒𝑑</sub> \= 0.3434, 𝑝 = 0.0631). In summary, we found that neural sensitivity to signal diagnosticity measured at change-consistent signals significantly correlated with individual subjects’ behavioral sensitivity to signal diagnosticity. By contrast, neural sensitivity to signal diagnosticity measured at change-inconsistent signals did not significantly correlate with behavioral sensitivity. The difference in correlation, however, was statistically significant in some (left IPS and left IFG) but not all brain regions within the frontoparietal network.”

      In the Methods section, we added on p.53:

      “Parametric and nonparametric tests for difference in correlation coefficients. We implemented both parametric and nonparametric tests to examine whether the difference in Pearson correlation coefficients was significant. In the parametric test, we used the Fisher 𝑧 transformation to transform the difference in correlation coefficients to the 𝑧 statistic. That is, for two correlation coefficients, 𝑟<sub>1</sub> (with sample size 𝑛<sub>2</sub>) and 𝑟<sub>2</sub>, (with sample size 𝑛<sub>1</sub>), the 𝑧 statistic of the difference in correlation is given by

      We referred to the correlation between neural and behavioral sensitivity at changeconsistent (blue balls) signals as 𝑟<sub>𝑏𝑙𝑢𝑒</sub>, and that at change-inconsistent (red balls) signals as 𝑟<sub>𝑟𝑒𝑑</sub>. For the Fisher 𝑧 transformation, 𝑟<sub>1</sub> \= 𝑟 𝑟<sub>𝑏𝑙𝑢𝑒</sub> and 𝑟<sub>2</sub> \= 𝑟<sub>𝑟𝑒𝑑</sub>. In the nonparametric test, we performed nonparametric bootstrapping to test for the difference in correlation (Efron & Tibshirani, 1994). That is, we resampled with replacement the dataset (subject-wise) and used the resampled dataset to compute the difference in correlation. We then repeated the above for 100,000 times so as to estimate the distribution of the difference in correlation coefficients, tested for significance and estimated p-value based on this distribution.”

      Another potential concern is that some important details about the parameter estimation for the system-neglect model are missing. In the respective section in the methods, the authors mention a nonlinear regression using Matlab's "fitnlm" function, but it remains unclear how the model was parameterized exactly. In particular, what are the properties of this nonlinear function, and what are the assumptions about the subject's motor noise? I could imagine that by using the inbuild function, the assumption was that residuals are Gaussian and homoscedastic, but it is possible that the assumption of homoscedasticity is violated, and residuals are systematically larger around p=0.5 compared to p=0 and p=1. Relatedly, in the parameter recovery analyses, the authors assume different levels of motor noise. Are these values representative of empirical values?

      We thank the reviewer for this excellent point. The reviewer touched on model parameterization, assumption of noise, and parameter recovery analysis. We answered these questions point-by-point below.

      On how our model was parameterized

      We parameterized the model according to the system-neglect model in Eq. (2) and estimated the alpha parameter separately for each level of transition probability and the beta parameter separately for each level of signal diagnosticity. As a result, we had a total of 6 parameters (3 alpha and 3 beta parameters) in the model. The system-neglect model is then called by fitnlm so that these parameters can be estimated. The term ‘nonlinear’ regression in fitnlm refers to the fact that you can specify any model (in our case the system-neglect model) and estimate its parameters when calling this function. In our use of fitnlm, we assume that the noise is Gaussian and homoscedastic (the default option).

      On the assumptions about subject’s motor noise

      We actually never called the noise ‘motor’ because it can be estimation noise as well. In the context of fitnlm, we assume that the noise is Gaussian and homoscedastic.

      On the possibility that homoscedasticity is violated

      We take the reviewer’s point. In response, we separately estimated the residual standard deviation at different probability intervals ([0.0–0.2), [0.2–0.4), [0.4–0.6), [0.6– 0.8), and [0.8–1.0]). The result is shown in the figure below. The black data points are the average residual standard deviation (across subjects) and the error bars are the standard error of the mean. The residual standard deviation is indeed heteroscedastic— smallest at 0.1 probability and increasing as probability increases and asymptote at 0.5 (Fig. S4).

      To examine how this would affect model fitting (parameter estimation), we performed parameter recovery analysis based on these empirically estimated, probabilitydependent residual standard deviation. That is, we simulated subjects’ probability estimates using the system-neglect model and added the heteroscedastic noise according to the empirical values and then estimated the parameter estimates of the system-neglect model. The recovered parameter estimates did not seem to be affected by the heteroscedasticity of the variance. The parameter recovery results were identical to the parameter recovery results when homoscedasticity was assumed. This suggested that although homoscedasticity was violated, it did not affect the accuracy of the parameter estimates (Fig.S4).

      We added a section ‘Impact of noise homoscedasticity on parameter estimation’ in Methods section (p.47-48) and a figure in the supplement (Fig. S4) to describe this:

      On whether the noise levels in parameter recovery analysis are representative of empirical values

      To address the reviewer’s question, we conducted a new analysis using maximum likelihood estimation to simultaneously estimate the system-neglect model and the noise level of each individual subject. To estimate each subject’s noise level, we incorporated a noise parameter into the system-neglect model. We assumed that probability estimates are noisy and modeled them with a Gaussian distribution where the noise parameter (𝜎,-./&) is the standard deviation. At each period, a probability estimate of regime shift was computed according to the system-neglect model where Θ is the set of parameters including parameters in the system-neglect model and the noise parameter. The likelihood function, 𝐿(Θ), is the probability of observing the subject’s actual probability estimate at period 𝑡, 𝑝), given Θ, 𝐿(Θ) = 𝑃(𝑝)|Θ). Since we modeled the noisy probability estimates with a Gaussian distribution, we can therefore express 𝐿(Θ) as 𝐿(Θ)~𝑁(𝑝); 𝑝)*+, 𝜎,-./&) where 𝑝)*+ is the probability estimate predicted by the system-neglect (SN) model at period 𝑡. As a reminder, we referred to a ‘period’ as the time when a new signal appeared during a trial (for a given transition probability and signal diagnosticity). To find that maximum likelihood estimates of ΘMLE, we summed over all periods the negative natural logarithm of likelihood and used MATLAB’s fmincon function to find ΘMLE. Across subjects, we found that the mean noise estimate was 0.1735 and ranged from 0.1118 to 0.2704 (Supplementary Figure S3).”

      Compared with our original parameter recovery analysis where the maximum noise level was set at 0.1, our data indicated that some subjects’ noise was larger than this value. Therefore, we expanded our parameter recovery analysis to include noise levels beyond 0.1 to up to 0.3. The results are now updated in Supplementary Fig. S3.

      We updated the parameter recovery section (p. 47) in Methods:

      The main study is based on N=30 subjects, as are the two control studies. Since this work is about individual differences (in particular w.r.t. to neural representations of noise and transition probabilities in the frontoparietal network and the vmPFC), I'm wondering how robust the results are. Is it likely that the results would replicate with a larger number of subjects? Can the two control studies be leveraged to address this concern to some extent?

      We can address the issue of robustness through looking at the effect size. In particular, with respect to individual differences in neural sensitivity of transition probability and signal diagnosticity, since the significant correlation coefficients between neural and behavioral sensitivity were between 0.4 and 0.58 for signal diagnosticity in frontoparietal network (Fig. 5C), and -0.38 and -0.37 for transition probability in vmPFC (Fig. 5D), the effect size of these correlation coefficients was considered medium to large (Cohen, 1992).

      It would be challenging to use the control studies to address the robustness concern. The two control studies did not allow us to examine individual differences – in particular with respect to neural selectivity of noise and transition probability – and therefore we think it is less likely to leverage the control studies. Having said that, it is possible to look at neural selectivity of noise (signal diagnosticity) in the first control experiment where subjects estimated the probability of blue regime in a task where there was no regime change (transition probability was 0). However, the fact that there were no regime shifts changed the nature of the task. Instead of always starting at the Red regime in the main experiment, in the first control experiment we randomly picked the regime to draw the signals from. It also changed the meaning and the dynamics of the signals (red and blue) that would appear. In the main experiment the blue signal is a signal consistent with change, but in the control experiment this is no longer the case. In the main experiment, the frequency of blue signals is contingent upon both noise and transition probability. In general, blue signals are less frequent than red signals because of small transition probabilities. But in the first control experiment, the frequency of blue signals may not be less frequent because the regime was blue in half of the trials. Due to these differences, we do not see how analyzing the control experiments could help in establishing robustness because we do not have a good prediction as to whether and how the neural selectivity would be impacted by these differences.

      It seems that the authors have not counterbalanced the colors and that subjects always reported the probability of the blue regime. If so, I'm wondering why this was not counterbalanced.

      We are aware of the reviewer’s concern. The first reason we did not do these (color counterbalancing and report blue/red regime balancing) was to not confuse the subjects in an already complicated task. Balancing these two variables also comes at the cost of sample size, which was the second reason we did not do it. Although we can elect to do these balancing at the between-subject level to not impact the task complexity, we could have introduced another confound that is the individual differences in how people respond to these variables. This is the third reason we were hesitant to do these counterbalancing.

      Reviewer #2 (Public review):

      Summary:

      This paper focuses on understanding the behavioral and neural basis of regime shift detection, a common yet hard problem that people encounter in an uncertain world.

      Using a regime-shift task, the authors examined cognitive factors influencing belief updates by manipulating signal diagnosticity and environmental volatility. Behaviorally, they have found that people demonstrate both over and under-reaction to changes given different combinations of task parameters, which can be explained by a unified system-neglect account. Neurally, the authors have found that the vmPFC-striatum network represents current belief as well as belief revision unique to the regime detection task. Meanwhile, the frontoparietal network represents cognitive factors influencing regime detection i.e., the strength of the evidence in support of the regime shift and the intertemporal belief probability. The authors further link behavioral signatures of system neglect with neural signals and have found dissociable patterns, with the frontoparietal network representing sensitivity to signal diagnosticity when the observation is consistent with regime shift and vmPFC representing environmental volatility, respectively. Together, these results shed light on the neural basis of regime shift detection especially the neural correlates of bias in belief update that can be observed behaviorally.

      Strengths:

      (1) The regime-shift detection task offers a solid ground to examine regime-shift detection without the potential confounding impact of learning and reward. Relatedly, the system-neglect modeling framework provides a unified account for both over or under-reacting to environmental changes, allowing researchers to extract a single parameter reflecting people's sensitivity to changes in decision variables and making it desirable for neuroimaging analysis to locate corresponding neural signals.

      Thank you for recognizing our task design and our system-neglect computational framework in understanding change detection.

      (2) The analysis for locating brain regions related to belief revision is solid. Within the current task, the authors look for brain regions whose activation covary with both current belief and belief change. Furthermore, the authors have ruled out the possibility of representing mere current belief or motor signal by comparing the current study results with two other studies. This set of analyses is very convincing.

      Thank you for recognizing our control studies in ruling out potential motor confounds in our neural findings on belief revision.

      (3) The section on using neuroimaging findings (i.e., the frontoparietal network is sensitive to evidence that signals regime shift) to reveal nuances in behavioral data (i.e., belief revision is more sensitive to evidence consistent with change) is very intriguing. I like how the authors structure the flow of the results, offering this as an extra piece of behavioral findings instead of ad-hoc implanting that into the computational modeling.

      Thank you for appreciating how we showed that neural insights can lead to new behavioral findings.

      Weaknesses:

      (1) The authors have presented two sets of neuroimaging results, and it is unclear to me how to reason between these two sets of results, especially for the frontoparietal network. On one hand, the frontoparietal network represents belief revision but not variables influencing belief revision (i.e., signal diagnosticity and environmental volatility). On the other hand, when it comes to understanding individual differences in regime detection, the frontoparietal network is associated with sensitivity to change and consistent evidence strength. I understand that belief revision correlates with sensitivity to signals, but it can probably benefit from formally discussing and connecting these two sets of results in discussion. Relatedly, the whole section on behavioral vs. neural slope results was not sufficiently discussed and connected to the existing literature in the discussion section. For example, the authors could provide more context to reason through the finding that striatum (but not vmPFC) is not sensitive to volatility.

      We thank the reviewer for the valuable suggestions.

      With regard to the first comment, we wish to clarify that we did not find frontoparietal network to represent belief revision. It was the vmPFC and ventral striatum that we found to represent belief revision (delta Pt in Fig. 3). For the frontoparietal network, we identified its involvement in our task through finding that its activity correlated with strength of change evidence (Fig. 4) and individual subjects’ sensitivity to signal diagnosticity (Fig. 5). Conceptually, these two findings reflect how individuals interpret the signals (signals consistent or inconsistent with change) in light of signal diagnosticity. This is because (1) strength of change evidence is defined as signals (+1 for signal consistent with change, and -1 for signal inconsistent with change) multiplied by signal diagnosticity and (2) sensitivity to signal diagnosticity reflects how individuals subjectively evaluate signal diagnosticity. At the theoretical level, these two findings can be interpreted through our computational framework in that both the strength of change evidence and sensitivity to signal diagnosticity contribute to estimating the likelihood of change (Eqs. 1 and 2). We added a paragraph in Discussion to talk about this.

      We added on p. 36:

      “For the frontoparietal network, we identified its involvement in our task through finding that its activity correlated with strength of change evidence (Fig. 4) and individual subjects’ sensitivity to signal diagnosticity (Fig. 5). Conceptually, these two findings reflect how individuals interpret the signals (signals consistent or inconsistent with change) in light of signal diagnosticity. This is because (1) strength of change evidence is defined as signals (+1 for signal consistent with change, and −1 for signal inconsistent with change) multiplied by signal diagnosticity and (2) sensitivity to signal diagnosticity reflects how individuals subjectively evaluate signal diagnosticity. At the theoretical level, these two findings can be interpreted through our computational framework in that both the strength of change evidence and sensitivity to signal diagnosticity contribute to estimating the likelihood of change (Equations 1 and 2 in Methods).”

      With regard to the second comment, we added a discussion on the behavioral and neural slope comparison. We pointed out previous papers conducting similar analysis (Vilares et al., 2011; Ting et al., 2015; Yang & Wu, 2020), their findings and how they relate to our results. Vilares et al. found that sensitivity to prior information (uncertainty in prior distribution) in the orbitofrontal cortex (OFC) and putamen correlated with behavioral measure of sensitivity to prior. In the current study, transition probability acts as prior in the system-neglect framework (Eq. 1) and we found that ventromedial prefrontal cortex represents subjects’ sensitivity to transition probability. Together, these results suggest that OFC (with vmPFC being part of OFC, see Wallis, 2011) is involved in the subjective evaluation of prior information in both static (Vilares et al., 2011) and dynamic environments (current study).

      We added on p. 37-38:

      “In the current study, our psychometric-neurometric analysis focused on comparing behavioral sensitivity with neural sensitivity to the system parameters (transition probability and signal diagnosticity). We measured sensitivity by estimating the slope of behavioral data (behavioral slope) and neural data (neural slope) in response to the system parameters. Previous studies had adopted a similar approach (Ting et al., 2015a; Vilares et al., 2012; Yang & Wu, 2020). For example, Vilares et al. (2012) found that sensitivity to prior information (uncertainty in prior distribution) in the orbitofrontal cortex (OFC) and putamen correlated with behavioral measure of sensitivity to the prior.

      In the current study, transition probability acts as prior in the system-neglect framework (Eq. 2 in Methods) and we found that ventromedial prefrontal cortex represents subjects’ sensitivity to transition probability. Together, these results suggest that OFC (with vmPFC being part of OFC, see Wallis, 2011) is involved in the subjective evaluation of prior information in both static (Vilares et al., 2012) and dynamic environments (current study). In addition, distinct from vmPFC in representing sensitivity to transition probability or prior, we found through the behavioral-neural slope comparison that the frontoparietal network represents how sensitive individual decision makers are to the diagnosticity of signals in revealing the true state (regime) of the environment.”

      (2) More details are needed for behavioral modeling under the system-neglect framework, particularly results on model comparison. I understand that this model has been validated in previous publications, but it is unclear to me whether it provides a superior model fit in the current dataset compared to other models (e.g., a model without \alpha or \beta). Relatedly, I wonder whether the final result section can be incorporated into modeling as well - i.e., the authors could test a variant of the model with two \betas depending on whether the observation is consistent with a regime shift and conduct model comparison.

      Thank you for the great suggestion. We rewrote the final Results section to specifically focus on model comparison. To address the reviewer’s suggestion (separately estimate beta parameters for change-consistent and change-inconsistent signals), we indeed found that these models were better than the original system-neglect model.

      To incorporate these new findings, we rewrote the entire final result section “Incorporating signal dependency into system-neglect model led to better models for regime-shift detection “(p.28-30).

      Recommendations for the authors:

      Reviewer #1 (Recommendations for the authors):

      (1) Use line numbers for the next round of reviews.

      We added line numbers in the revised manuscript.

      (2) Figure 2b: Can the empirical results be reproduced by the system-neglect model? This would complement the analyses presented in Figure S4.

      Yes. We now add Figure S6 based on system-neglect model fits. For each subject, we first computed period-by-period probability estimates based on the parameter estimates of the system-neglect model. Second, we computed index of overreaction (IO) for each combination of transition probability and signal diagnosticity. Third, we plot the IO like we did using empirical results in Fig. 2b. We found that the empirical results in Fig. 2b are similar to the system-neglect model shown in Figure S6, indicating that the empirical results can be reproduced by the model.

      (3) Page 14: Instead of referring to the "Methods" in general, you could be more specific about where the relevant information can be found.

      Fixed. We changed “See Methods” to “See System-neglect model in Methods”.

      (4) Page 18: Consider avoiding the term "more significantly". Consider effect sizes if interested in comparing effects to each other.

      Fixed. On page 19, we changed that to

      “In the second analysis, we found that for both vmPFC and ventral striatum, the regression coefficient of 𝑃) was significantly different between Experiment 1 and Experiment 2 (Fig. 3C) and between Experiment 1 and Experiment 3 (Fig. 3D; also see Tables S5 and S6 in SI).”

      (5) Page 30: Cite key studies using reversal-learning paradigms. Currently, readers less familiar with the literature might have difficulties with this.

      We now cite key studies using reversal-learning paradigms on p.32:

      “Our work is closely related to the reversal-learning paradigm—the standard paradigm in neuroscience and psychology to study change detection (Fellows & Farah, 2003; Izquierdo et al., 2017; O'Doherty et al., 2001; Schoenbaum et al., 2000; Walton et al., 2010). In a typical reversal-learning task, human or animal subjects choose between two options that differ in the reward magnitude or probability of receiving a reward. Through reward feedback the participants gradually learn the reward contingencies associated with the options and have to update knowledge about reward contingencies when contingencies are switched in order to maximize rewards.”

      Reviewer #2 (Recommendations for the authors):

      (1) Some literature on change detection seems missing. For example, the author should also cite Muller, T. H., Mars, R. B., Behrens, T. E., & O'Reilly, J. X. (2019). Control of entropy in neural models of environmental state. elife, 8, e39404. This paper suggests that medial PFC is correlated with the entropy of the current state, which is closely related to regime change and environmental volatility.

      Thank you for pointing to this paper. We have now added it and other related papers in the Introduction and Discussion.

      In Introduction, we added on p.5-6:

      “Different behavioral paradigms, most notably reversal learning, and computational models were developed to investigate its neurocomputational substrates (Behrens et al., 2007; Izquierdo et al., 2017; Payzan-LeNestour et al., 2011, 2013; Nasser et al., 2010; McGuire et al., 2014; Muller et al., 2019). Key findings on the neural implementations for such learning include identifying brain areas and networks that track volatility in the environment (rate of change) (Behrens et al., 2007), the uncertainty or entropy of the current state of the environment (Muller et al., 2019), participants’ beliefs about change (Payzan-LeNestour et al., 2011; McGuire et al., 2014; Kao et al., 2020), and their uncertainty about whether a change had occurred (McGuire et al., 2014; Kao et al., 2020).”

      In Discussion (p.35), we added a new paragraph:

      “Related to OFC function in decision making and reinforcement learning, Wilson et al. (2014) proposed that OFC is involved in inferring the current state of the environment. For example, medial OFC had been shown to represent probability distribution on possible states of the environment (Chan et al., 2016), the current task state (Schuck et al., 2016) and uncertainty or entropy associated with the state of the environment (Muller et al., 2019). In the context of regime-shift detection, regimes can be regarded as states of the environment and therefore a change in regime indicates a change in the state of the environment. Muller et al. (2019) found that in dynamic environments where changes in the state of the environment happen regularly, medial OFC represented the level of uncertainty in the current state of the environment. Our finding that vmPFC represented individual participants’ probability estimates of regime shifts suggest that vmPFC and/or OFC are involved in inferring the current state of the environment through estimating whether the state has changed. Our finding that vmPFC represented individual participants’ sensitivity to transition probability further suggest that vmPFC and/or OFC contribute to individual participants’ biases in state inference (over- and underreactions to change) in how these brain areas respond to the volatility of the environment.”

      (2) The language used when describing the selective relationship between frontoparietal network activation and change-consistent signal can be clearer. When describing separating those two signals, the authors refer to them as when the 'blue' signal shows up and when the 'red' signal shows up, assuming that the current belief state is blue. This is a little confusing cuz it is hard to keep in mind what is the default color in this example. It would be more intuitive if the author used language such as the 'change consistent' signal.

      Thank you for the suggestion. We have changed the wording according to your suggestion. That is, we say ‘change-consistent (blue) signals’ and ‘change-inconsistent (red) signals’ throughout pages 22-28.

      (3) Figure 4B highlights dmPFC. However, in the associated text, it says p = .10 so it is not significant. To avoid misleading readers, I would recommend pointing this out explicitly beyond saying 'most brain regions in the frontoparietal network also correlated with the intertemporal prior'.

      Thank you for pointing this out. We now say on p.20

      “With independent (leave-one-subject-out, LOSO) ROI analysis, we examined whether brain regions in the frontoparietal network (shown to represent strength of change evidence) correlated with intertemporal prior and found that all brain regions, with the exception of dmPFC, in the frontoparietal network correlated with the intertemporal prior.”

      (4) There is a full paragraph in the discussion talking about the central opercular cortex, but this terminology has not shown up in the main body of the paper. If this is an important brain region to the authors, I would recommend mentioning it more often in the result section.

      Thank you for this suggestion. We have now added central opercular cortex in the Results section (p.18):

      “For 𝑃<sub>𝑡</sub>, we found that the ventromedial prefrontal cortex (vmPFC) and ventral striatum correlated with this behavioral measure of subjects’ belief about change. In addition, many other brain regions, including the motor cortex, central opercular cortex, insula, occipital cortex, and the cerebellum also significantly correlated with 𝑃<sub>𝑡</sub>.”

      (5) The authors have claimed that people make more extreme estimates under high diagnosticity (Supplementary Figure 1). This is an interesting point because it seems to be different from what is shown in the main graph where it seems that people are not extreme enough compared to an ideal Bayesian observer. I understand that these are effects being investigated under different circumstances. It would be helpful if for Supplementary Figure 1 the authors could overlay, or generate a different figure showing what an ideal Bayesian observer would do in this situation.

      We thank the reviewer for pointing this out. We wish to clarify that when we said “more extreme estimates under high diagnosticity” we meant compared with low diagnosticity and not with the ideal Bayesian observer. We clarified this point by rephrasing our sentence on p.11:

      “We also found that subjects tended to give more extreme Pt under high signal diagnosticity than low diagnosticity (Fig. S1 in Supplementary Information, SI).”

      When it comes to comparing subjects’ probability estimates with the normative Bayesian, subjects tended to “underreact” under high diagnosticity. This can be seen in Fig. 4B, which shows a trend of increasing underreaction (or decreasing overreaction) as diagnosticity increased (row-wise comparison for a given transition probability).

      We see the reviewer’s point in overlaying the Bayesian on Fig. S1 and update it by adding the normative Bayesian in orange.

    1. More work does not necessarily mean a student is learning or is engaged. Students who are gifted can begin to feel their giftedness is a punishment if they are consistently asked or required to complete a greater quantity of work. Choosing quality opportunities that match the interest of students will genuinely be of benefit to students who are gifted.

      Quality and depth, not quantity, determine whether learning is meaningful to gifted students.

    2. the possibility of failure can lead to some unusual and distracting emotional situations.

      What specific scaffolds or affirmations help gifted students navigate a class (like music) where mistakes are part of the learning process?

    3. A student with an IQ of 150, however, is often in an inclusion situation with no services or accommodations. Their IQs are equally different from those students who are considered average (IQ 100); however, the attention paid to their educational needs is vastly unequal (Silverman, 1993). Webb et al. (1994) explain: “Gifted children are not simply decorated normal children—they are, indeed, fundamentally different. A child with IQ 145 is as different from the normal IQ of 100 as the child of IQ 55” (p. 31).

      Both ends of the bell curve require differentiated instruction, but unfortunately only one typically receives it.

    4. Through the standardization process

      I didn’t realise gifted education came from the same tests originally meant to exclude kids. It shows how definitions of ability can shift across time.

    5. The differences and disabilities of students who are intellectually gifted are often delayed, ignored, and denied.

      Instead of receiving needed academic challenge, gifted students frequently slip through the cracks because teachers assume they will be “fine” without support.

    6. They are also at great risk in our classrooms, which are often designed for the average student and to offer accommodations for students with other types of differences and disabilities.

      This is harsh truth that many of us don't think about because gifted kids are never thought of as a vulnerable group. The idea that not being challenged can actually harm their motivation and development is surprising and concerning.

    1. TheNational Park Service offered apatriotic story of the location andera, complete with re-creationsof the reading of the Declara-tion of Independence from 1776and debates over the Constitu-tion in 1787. Skilled interpreterswearing historic costumes illu-minated the time and place andbroadened the audience’s under-standing of the experiences. Thestatement by Daniel Webster in1837 emblazoned on the wall ofthe National Constitution Center,“One country, one Constitution,one destiny,” celebrates the crea-tion story of the United Statesof America

      Launius says that places like Independence Hall are usually presented in a positive, patriotic way. They highlight the greatness of America’s founding rather than showing the harder or more uncomfortable parts of history.

    1. We hope with this you can be a more informed user of social media, better able to participate, protect yourself, and make it a valuable experience for you and others you interact with.

      I believe this is the key idea of writing this book: the ultimate purpose is to inform others, even those who haven't read it. The author is actually using the idea of viral communication on the internet. As long as a group of people on social media is educated and agrees on the book, they will spread the book's idea to others in the future no matter unintended or intended.

    1. two monumental stoneware vessels that were made by David Drake (also known as Dave the Potter), an enslaved potter and poet from Old Edgefield, South Carolina.

      The museum is discussing two large pottery pieces created by David Drake. He was an enslaved man who made pottery and wrote short poems. This is important because enslaved people were usually not allowed to learn to read or write, so his work is rare and meaningful.

    1. “It’s easily the most controversial statue I’ve ever made,” he observes.

      Frudakis admits that this artwork has become a symbol of conflict. By calling it “the most controversial,” he acknowledges that Rizzo’s legacy is not neutral and that the statue evokes strong emotions, both positive and negative, within the community.

    1. History "offers the only extensive evidential base for the contemplation and analysis of how societies function, and people need to have some sense of how societies function simply to run their own lives."

      Stearns is saying that history gives us the best information for understanding how societies work. If we don’t know how things worked in the past, it’s hard to understand how things work today. This matters for regular people too, not just historians, because it helps us make smarter decisions in everyday life.