20 Matching Annotations
  1. Mar 2026
    1. This visualization is difficult for a casual viewer to decipher,and this is necessarily so, because the visualization is itself aresearch tool. It is a way of analyzing information more than afinished representation of the conclusion.

      Exactly, for someone with a keen eye for collecting data, they probably see this and they are in complete data collection bliss.

    2. allows us to merge things createdat dramatically different times to create what are in effect newmodern images which potentially reveal things about the pastthat the original artifacts did not.11

      That is quite fascinating. This is a good example of what I was just talking about with new technologies opening up for new methods that did not exist from back then they did not exist.

    3. This brings me to new technologies, which are by no meansuncontroversial. Historians like myself who advocate the use ofthese technologies can be accused of having deserted the forces oflight and embraced the forces of reaction.

      Ultimately, if something makes something easier, especially something as complex as measuring space, I think it is totally fine to embrace new technologies in this way.

    4. They are not separate from spatial practice because in largemeasure they are what guide the human labor that creates kitchens,bathrooms, living rooms, roads, train stations, airports, air trafficcontrol and entire landscapes.

      This is fascinating since it is making me truly appreciate how much the chronicling of space and time are so heavily intertwined. This is correct, you cannot measure one without the other.

    1. In other words, we need smart big data because it can not only adequately represent a sufficient number of relevant features of humanistic objects of inquiry to enable the level of precision and nuance scholars in the humanities need, but it can also provide us with a sufficient amount of data to enable quantitative methods of inquiry that help us transgress the limitations inherent in methods based on close reading strategies.

      I like this ending statement that we need to find a balance of both smart and big data. It makes sense with how essential both of them are to DH.

    2. Basically, there are two possible ways to do this: one is automatic annotation, the other is crowdsourcing. Automation refers to various heuristics of discovering implicit units, structures, patterns and relations, and of making them explicit in the data. Crowdsourcing, on the other hand, relies on breaking down a large task into such small units that each of these little tasks can be performed in a distributed way by a large number of volunteers.

      This is interesting. While the automation seems like it gets things done much more quickly, while still finding patterns. However, the crowdsourcing seems much more effective if you want effective research to be done, while still being relatively quick due to it being done by individual small units.

    3. Even without using a new term, we can now redefine what we mean by data in the humanities. Data in the humanities could be considered a digital, selectively constructed, machine-actionable abstraction representing some aspects of a given object of humanistic inquiry.

      See, I like this definition of data significantly more than it just being that source. It being separated as a distinct digital replication of a source works much better for me.

    4. But rarely would they consider their objects of study to be “data.” However, in the humanities just as in other areas of research, we are increasingly dealing with “data.”

      I like this hesitation to just dismiss all of these individual sources as just being "data." This is because making these things amount to just being "data" rips away all of the purpose, meaning, and impact of each individual source.

  2. Jan 2026
    1. All three of the areas I’ve described here - multimodal generation, Deep Research, and handwriting transcription - have changed dramatically in the past twelve months.

      The rapidness of how much better it is getting is really scary. I do not think anyone predicted just how much it would evolve last year. This just means the advancements will keep getting faster and faster!

    2. This year also saw the rise of reasoning and agentic models. To oversimplify, reasoning models break their responses down into smaller sequential steps (called “chain of thought”) and spend more time and computing power on generating an answer.

      This is really scary! Full-blown Artificial Intelligence is not that far away! I really cannot believe we are in this timeline. Pop culture was warning about this for decades!

    3. Any historian interested in public engagement needs to pay attention to this development.

      Absolutely. With the rise of this new technology has brought with it an ever-increasing need to ensure that history is actually preserved, and not a false one.

    4. Whether or not you guessed correctly, this exercise illustrates a larger point: generative AI has gotten really good, really fast, at generating multimodal content, or stuff that isn’t just text (i.e. images, video, and audio).

      Wow, I thought I guessed correctly as to which one was AI. I could have sworn it was the second one! But nope, I was wrong. Very scary!

    1. We hope this overview helps explain why digitization within archival institutions proceeds the way it does – and why we may never, in fact, digitize everything.

      It genuinely did, I wondered the same thing at one point, but seeing the process that goes behind every single digital archive makes me see why.

    2. Because digitization involves an investment of time and resources, we need to make sure we get it right – that the electronic files we produce are adequately representing the archival originals. That means our process will need to incorporate quality control checks.

      With the amount of time that it takes to digitize something, I can see why they would not want to do it with EVERYTHING. However, it does pay off by making it accessible!

    3. At best it results in a digital “surrogate,” an approximation (even if a very good one) of a dimension of the record.

      Reading this it can be clear why someone may be disappointed with a digital record, since it is technically not the original record. However, it still has all of the same information, so I don't see it as that big of a downgrade.

    4. As archivists we like these questions because they tell us that people are eager for access to archival records. They also show that people realize that not everything is digitized. Indeed only a tiny fraction of the world’s primary resources are available digitally.

      Sure, some individuals may be more eager for physical records, but it should not be a question that digital archives are significantly easier to access. So I think that is a big factor to consider.

    1. ne development far outstripped the impact of any other during the 1990s. This was the arrival of the Internet, but more especially the World Wide Web. The first graphical browser, Mosaic, appeared on the scene in 1993. Now the use of the Internet is a vital part of any academic activity. A generation of students has grown up with it and naturally looks to it as the first source of any information.

      This is really where a lot of that data entry that was done before really began to pay off. Now this data could easily be accessed by everyone and anyone using the Internet, giving rise to the digital humanities we know today.

    2. If any single-word term can be used to describe this period, it would almost certainly be "consolidation." More people were using methodologies developed during the early period. More electronic texts were being created and more projects using the same applications were started. Knowledge of what is possible had gradually spread through normal scholarly channels of communication, and more and more people had come across computers in their everyday life and had begun to think about what computers might do for their research and teaching.

      This would be on the path to getting to GUI, so I think "consolidation" is a good term for this period. As well as making data entry significantly more accessible to your average person.

    3. The other widely used citation scheme was more dependent on punched card format. In this scheme, often called "fixed format", every line began with a coded sequence of characters giving citation information. Each unit within the citation was positioned in specific columns across the line, for example the title in columns 1–3, verse number in columns 5–6, and line number in columns 7–9. The entry of this information was speeded up by functions on the punched card machine, but the information also occupied more space within the computer file.

      Like I said before, this feels so cryptic and difficult in comparison to what we have today.

    4. At this time much attention was paid to the limitations of the technology. Data to be analyzed were either texts or numbers. They were input laboriously by hand either on punched cards, with each card holding up to eighty characters or one line of text (uppercase letters only), or on paper tape, where lower-case letters were perhaps possible but which could not be read in any way at all by a human being.

      As someone who was born in an age where the graphic user interface (GUI) was the standard when it came to Computers, it is always really interesting to hear about this time in history. Inputing all of that data sounds like such a painstaking process for the time.