243 Matching Annotations
  1. Apr 2026
    1. Claude Opus 4.7 demonstrates strong substantive accuracy on BigLaw Bench for Harvey, scoring 90.9% at high effort with better reasoning calibration on review tables and noticeably smarter handling of ambiguous document editing tasks.

      在法律文档处理中达到90.9%的准确率,特别是在处理模糊文档编辑任务时的智能提升,展示了AI在专业领域的深度应用能力,这种进步将极大扩展AI在法律和合规领域的应用价值。

    2. Opus 4.7 uses an updated tokenizer that improves how the model processes text. The tradeoff is that the same input can map to more tokens—roughly 1.0–1.35× depending on the content type.

      tokenizer的更新虽然增加了token使用量,但提高了文本处理效率,这反映了AI模型在基础架构上的持续优化,这种优化虽然带来短期成本增加,但长期将提升AI的处理能力和准确性。

    1. Codex just hit 3 million weekly active users, our APIs process more than 15 billion tokens per minute, and GPT‑5.4 is driving record engagement across agentic workflows.

      这些惊人的使用指标展示了AI技术在实际应用中的大规模采用。特别是每分钟处理150亿个token的能力,反映了企业对AI处理能力的巨大需求,以及AI已经从实验阶段进入实际工作流程的临界点。

    1. Add screenshot-based LLM judge evaluator, screenshot collector, and --parallelize flag

      引入基于截图的LLM评估器和并行化功能是一个令人惊讶的创新。通过截图评估AI模型的性能,可以更直观地理解自动化过程中的视觉理解能力,而并行化功能则大大提高了基准测试的效率,这代表了AI系统评估方法的重要进步。

    1. Claude code 可以并行 12个 subagent,几分钟,20x 的限额就到了

      令人惊讶的是:Claude code的并发处理能力如此强大,能够同时运行12个子代理,但同时也暴露了其API使用限制的脆弱性,几分钟内就达到20倍的限额,这表明即使是高级AI模型也存在明显的使用边界,可能影响大规模应用场景。

    1. Within a few months, they have more than a dozen production enterprise deployments & are processing over a billion events per hour.

      令人惊讶的是:Artemis安全公司在短短几个月内就处理了每小时超过10亿个安全事件,这种数据处理规模反映了现代企业面临的网络安全威胁的惊人频率和复杂性。

    1. In order to use the relative entropy method, we reconstruct the numerical solution via tensor-product Smoothness-Increasing Accuracy-Conserving (SIAC) filtering which has superconvergence properties.

      SIAC 滤波器的「超收敛」性质令人印象深刻:对多项式次数为 q 的 DG 解进行 SIAC 后处理后,收敛阶从 q+1 跃升至 2q+1——精度几乎翻倍,却几乎不增加计算代价。这是数值分析中罕见的「免费午餐」:滤波本身是线性操作,计算量微乎其微,却能将误差的收敛速率提升一个整量级。

    1. These representations track the operative emotion concept at a given token position in a conversation, activating in accordance with that emotion's relevance to processing the present context and predicting upcoming text.

      「在特定 token 位置追踪当前生效的情绪概念」——这句话揭示了一个深刻洞见:情绪不是持续状态,而是逐词涌现的动态标注。这与人类神经科学中「情绪是对当前感知的实时评估」高度吻合,暗示 LLM 在没有神经元的情况下,重演了大脑皮层处理情绪的某种计算逻辑。

    1. It achieves state-of-the-art performance on streaming benchmarks and supports a real-time demo system with ASR and TTS running at 2 FPS on two 80G accelerators

      大多数人认为实时视频处理需要极高的计算资源和帧率才能有效,但作者仅用两块80G加速器就实现了2 FPS的实时系统,并达到了最先进的性能。这一结果挑战了高性能视频处理需要大量计算资源的共识,暗示通过优化算法和架构可以显著降低实时视频处理的计算门槛。

    2. current approaches often rely on decoupled trigger-response pipelines or are limited to captioning-style narration, reducing their effectiveness for open-ended question answering and long-horizon interaction

      大多数人认为现有的视频大模型可以通过简单的触发-响应管道或描述式叙述来处理实时视频流,但作者认为这种方法对于开放式问答和长时程交互效果有限。这是一个反直觉的观点,因为它挑战了当前视频处理领域的常规做法,暗示需要更集成的端到端方法来真正实现实时视频理解。

    1. Rather than treating a complex document as a single monolithic task, Deep Extract deploys sub-agents to break it down and conquer each piece, which is what allows it to remain accurate even on documents with thousands of rows across hundreds of pages.

      大多数人可能认为处理复杂文档的最佳方式是将其作为一个整体来处理,保持上下文完整性。但作者提出将复杂文档分解为多个子任务并由子代理分别处理的方法更有效,这一方法挑战了文档处理中'整体优于部分'的传统认知,暗示分解策略可能更适合处理超长文档。

    1. By applying compute otherwise that goes unutilized to predict and verify additional tokens in parallel (up to three in this implementation), throughput at high interactivity is increased.

      大多数人认为计算资源应该用于当前任务,但作者提出利用未充分利用的计算资源并行预测额外令牌的创新方法,这挑战了传统计算资源分配的常识,暗示了AI计算效率的全新可能性。

  2. Dec 2025
  3. Oct 2025
    1. From a memory perspective, sleep can be understood as critically important for normal memory function, given the lasting ramifications of consolidation.

      for - key insight - paraphrase - adjacency - memory consolidation - sleep - massive unconscious parallel processing - From a memory perspective, - sleep can be understood as critically important for normal memory function, - given the lasting ramifications of consolidation. - Consolidation is the establishment of new connections - anchoring recent memories within relevant knowledge networks - While consolidation happens, some conscious experience (the dream) may be synthesized as the memory processing unfolds - Dreams reflect a storyline generated to make sense of a subset of activated memory fragments. - Consolidation that wires new connections happens across the entire cerebral context, without the constraints that come with conscious experience. - Unconscious processing during sleep takes advantage of massive parallel processing to connect all these thoughts together. - Dreams reflect a small portion of overnight memory consolidation work.

  4. Aug 2025
    1. this system that actually tells you which one is you which one is not you is the self is not the self yeah so the immune system

      for - adjacency - brain - identity - immune system - immune system involved with identity at a microscopic level - which molecule is part of "you"? - which moleculev is NOT part of you - immune system preceded neural system

      ? - maybe there is similarity between neural processing and immune system response?

  5. May 2025
    1. perceived by oneself “in here.” In this sense, the world consists of objects outthere in space (the container that holds them) before me as the perceivingsubject.

      for - adjacency - Indyweb dev - natural language - timebinding - parallel vs serial processing - comparison - spoken vs written language - what's also interesting is that spoken language is timebinding, sequential and our written language descended from that, - in spite of written language existing in 2D and 3D space, it inherited sequential flow, even though it does not have to - In this sense, legacy spoken language system constrains written language to be - serial - sequential and - timebound instead of - parallel - Read any written text and you will observe that the pattern is sequential - We constrain our syntax it to "flow" sequentially in 2D space, even though there is absolutely no other reason to constrain it to do so - This also reveals another implicit rule about language, that it assumes we can only focus our attention on one aspect of reality at a time

  6. Nov 2024
    1. Een chunk (letterlijk ‘brok’) is een verzameling elementen die sterke associaties met elkaar hebben. Samen vormen ze een betekenisvolle informatie-eenheid. Die chunks, groot of klein, gebruiken we in ons interne informatieverwerkings- en geheugensysteem. Ons brein houdt namelijk van logica en voorspelbare patronen. Het opdelen van informatie gebeurt automatisch en continu, maar kan ook bewust worden ingezet. Dat heet doel-georiënteerde chunking.Ons brein kan slechts een aantal zaken opslaan in het kortetermijngeheugen. Maar door veel gegevens te groeperen in kleinere brokjes informatie, kunnen we de limieten van ons geheugen uitdagen. En dus meer informatie verwerken en onthouden.

      Chapeau! Een Belgische website kaart dit aan in de context gezond leven.

  7. Oct 2024
    1. Engagingwith the slip box should feel exciting, not anxiety-producing.

      I often find that people who discuss "workflows" and the idea of "processing" their notes are the ones who are falling trap to the anxiety-producing side of the work.

      BD should have found more exciting words for "processing" which he uses two more times in the next paragraph.

      This relates to Luhmann's quote about only doing what is easy/fun/flow:<br /> - https://hypothes.is/a/TQyC1q1HEe2J9fOtlKPXmA<br /> - https://hypothes.is/a/EyKrfK1WEe2RpEuwUuFA7A

      Compare: - being trapped in the box: https://hypothes.is/a/AY7ABO0qEeympasqOZHoMQ - idea of drudgery in the phrase "word processing"

  8. Aug 2024
    1. Typewriter Video Series - Episode 147: Font Sizes and the Writing Process by [[Joe Van Cleave]]

      typewriters for note making

      double or 1 1/2 spacing with smaller typefaces may be more efficient for drafting documents, especially first drafts

      editing on actual paper can be more useful for some

      Drafting on a full sheet folded in half provides a book-like reading experience for reading/editing and provides an automatic backing sheet

      typewritten (or printed) sheets may be easier to see and revise than digital formats which may hide text the way ancient scrolls did for those who read them.

      Jack Kerouac used rolls of paper to provide continuous writing experience. Doesn't waste the margins of paper at the top/bottom. This may be very useful for first drafts.

      JVC likes to thread rolls of paper into typewriters opposite to the original curl so as to flatten the paper out in the end.

    1. For true deep processing and learning, intellectualism, one must think beyond the single source they are consuming and think about everything they know. Although keep in mind selective attention for true learning and thinking.

      This process is habitualized by means of Zettelkasten and further aided in tool like hypothes.is

  9. Jul 2024
    1. Whoosh provides methods for computing the “key terms” of a set of documents. For these methods, “key terms” basically means terms that are frequent in the given documents, but relatively infrequent in the indexed collection as a whole.

      Very interesting method, and way of looking at the signal. "What makes a document exceptional because something is common within itself and uncommon without".

  10. Jun 2024
    1. Testing culture also discourages deep reading, critics say, because it emphasizes close reading of excerpts, for example, to study a particular literary technique, rather than reading entire works.

      Indeed. But testing in general, as it is done currently, in modern formal education, discourages deep learning as opposed to shallow learning.

      Why? Because tests with marks implore students to start learning at max 3 days before the test, thus getting knowledge into short-term memory and not long term memory. Rendering the process of learning virtually useless even though they "pass" the curriculum.

      I know this because I was such a student, and saw it all around me with virtually every other student I met, and I was in HAVO, a level not considered "low".

      It does not help that teachers, or the system, expect students to know how to learn (efficiently) without it ever being taught to them.

      My message to the system: start teaching students how to learn the moment they enter high school

  11. May 2024
    1. Matthew van der Hoorn Yes totally agree but could be used for creating a draft to work with, that's always the angle I try to take buy hear what you are saying Matthew!

      Reply to Nidhi Sachdeva: Nidhi Sachdeva, PhD Just went through the micro-lesson itself. In the context of teachers using to generate instruction examples, I do not argue against that. The teacher does not have to learn the content, or so I hope.

      However, I would argue that the learners themselves should try to come up with examples or analogies, etc. But this depends on the learner's learning skills, which should be taught in schools in the first place.

    2. ***Deep Processing***-> It's important in learning. It's when our brain constructs meaning and says, "Ah, I get it, this makes sense." -> It's when new knowledge establishes connections to your pre-existing knowledge.-> When done well, It's what makes the knowledge easily retrievable when you need it. How do we achieve deep processing in learning? 👉🏽 STORIES, EXPLANATIONS, EXAMPLES, ANALOGIES and more - they all promote deep meaningful processing. 🤔BUT, it's not always easy to come up with stories and examples. It's also time-consuming. You can ask you AI buddies to help with that. We have it now, let's leverage it. Here's a microlesson developed on 7taps Microlearning about this topic.

      Reply to Nidhi Sachdeva: I agree mostly, but I would advice against using AI for this. If your brain is not doing the work (the AI is coming up with the story/analogy) it is much less effective. Dr. Sönke Ahrens already said: "He who does the effort, does the learning."

      I would bet that Cognitive Load Theory also would show that there is much less optimized intrinsic cognitive load (load stemming from the building or automation of cognitive schemas) when another person, or the AI, is thinking of the analogies.


      https://www.linkedin.com/feed/update/urn:li:activity:7199396764536221698/

  12. Apr 2024
  13. Mar 2024
    1. When processing an item in your in list the first question you need to ask is: is it actionable?—in other words, do you need to do something? If the answer is NO, you either throw it away if you no longer need it, keep it as reference material (“I will probably need this article again some day…”), add it to a some day/maybe list (for things like “learn Indonesian”), or incubate it. Wait, what‽ Sit on it? Yes, sort of. If it’s something that you want to remind yourself about later (“I really didn’t understand this article, I should have a look at it again in two weeks”) it should go into your calendar or your tickler file which will soon be explained. (Yes, even the weird name.)

      First, ask yourself if the item is actionable. Then, series of stuff you might do: throw away, reference, someday/maybe, incubate (calendar/tickler)

    1. Samuel Hartlib was well aware of this improvement. While extolling the clever invention of Harrison, Hartlib noted that combinations and links con-stituted the ‘argumentative part’ of the card index.60

      Hartlib Papers 30/4/47A, Ephemerides 1640, Part 2.

      In extolling the Ark of Studies created by Thomas Harrison, Samuel Hartlib indicated that the combinations of information and the potential links between them created the "argumentative part" of the system. In some sense this seems to be analogous to the the processing power of an information system if not specifically creating its consciousness.

    1. some of our older applications rely substantially on manual extract, transform and load (ETL)processes to pass data from one system to another. This substantially increases the volumeof customer and staff data in transit on the network, which in a modern data managementand reporting infrastructure would be encapsulated in secure, automated end-to-end

      Reliance on ETL seen as risky

      I’m not convinced about this. Real-time API connectivity between systems is a great goal…very responsive to changes filtering through disparate systems. But a lot of “modern” processing is still done by ETL batches (sometimes daily, sometimes hourly, sometimes every minute).

  14. Feb 2024
  15. Jan 2024
    1. Deep processing is the foundation of all learning. It refers to your ability to think about information critically, find relationships, make sense of new information, and organise it into meaningful knowledge in your memory.
    1. I am particularly interested in how performance style and expressive vocabulary changes over time, as evidenced on sound recordings. I enjoy exploring aesthetics questions both empirically through experiments and measurements as well as philosophically, i.e. in their historical and cultural context.I try to embrace interdisciplinary approaches (e.g. cognitive neuroscience and perception as well as ethnographic and archival work) and learn from cross cultural investigations. I particularly like working with performers who are interested in research.
  16. Nov 2023

    Tags

    Annotators

  17. Sep 2023
    1. (1:20.00-1:40.00) What he describes is the following: Most of his notes originate from the digital using hypothes.is, where he reads material online and can annotate, highlight, and tag to help future him find the material by tag or bulk digital search. He calls his hypothes.is a commonplace book that is somewhat pre-organized.

      Aldrich continues by explaining that in his commonplace hypothes.is his notes are not interlinked in a Luhmannian Zettelkasten sense, but he "sucks the data" right into Obsidian where he plays around with the content and does some of that interlinking and massage it.

      Then, the best of the best material, or that which he is most interested in working with, writing about, etc., converted into a more Luhmannesque type Zettelkasten where it is much more densely interlinked. He emphasizes that his Luhmann zettelkasten is mostly consisting of his own thoughts and is very well-developed, to the point where he can "take a string of 20 cards and ostensibly it's its own essay and then publish it as a blog post or article."

  18. Aug 2023
    1. If there’s only an asterisk: Click the style name, then move the pointer over the style name in the Paragraph Styles pop-up menu. Click the arrow that appears, then choose Redefine from Selection.

      Pages is so much more impressive than you'd expect in so many ways, but damn...

      The way styles are handled still perplexes the shit out of me... even after consuming this document.

    1. The essence for this video is correct; active learning, progressive summarization, deep processing, relational analytical thinking, even evaluative.

      Yet, the implementation is severely lacking; marginalia, text writing, etc.

      Better would be the use of mindmaps or GRINDEmaps. I personally would combine it with the Antinet of course.

      I do like this guy's teaching style though 😂

  19. Jul 2023
    1. Code for processing data samples can get messy and hard to maintain; we ideally want our dataset code to be decoupled from our model training code for better readability and modularity.

      Code for data processing and model training should be separated as different modules.

    1. We prioritize what we see versus what we hear, why is that? Now, what comes to mind when I say that is when, somebody is saying no, but shaking their head yes. And so we have this disconnect, but we tend to prioritize what the action and not what we're hearing. So something that we visually see instead of what we hear.Speaker 1There isn't a definitive answer on that, but one source of insight on why do we do that, it could be related to the neurological real estate that's taken up by our visual experience. There's far more of our cortex, the outer layer of our brain that responds to visual information than any other form of information

      (13:36) Perhaps this is also why visual information is so useful for learning and cognition (see GRINDE)... Maybe the visual medium should be used more in instruction instead of primarily auditory lectures (do take into account redundancy and other medium effects from CLT though)

  20. Jun 2023
    1. When it comes to thinking, the Zettelkasten solves an important issue which is the problem of scope, which is impossible at the current moment in mindmapping software such as Concepts.

      Mainly, Zettelkasten allows you gain a birds-eye holistic view of a topic, branch, or line of thought, while allowing you to at the same time also gain a microscopic view of an "atomic" idea within that thought-stream, therefore creating virtually infinite zoom-in and zoom-out capability. This is very, very, beneficial to the process of deep thinking and intellectual work.

    1. Recent work in computer vision has shown that common im-age datasets contain a non-trivial amount of near-duplicateimages. For instance CIFAR-10 has 3.3% overlap betweentrain and test images (Barz & Denzler, 2019). This results inan over-reporting of the generalization performance of ma-chine learning systems.

      CIFAR-10 performance results are overestimates since some of the training data is essentially in the test set.

  21. Mar 2023
    1. In short, in the absence of legal tender laws, the seller will not accept anything but money of certain value (good money), but the existence of legal tender laws will cause the buyer to offer only money with the lowest commodity value (bad money), as the creditor must accept such money at face value.

      During the coronavirus pandemic, many vendors facing inflation began to pass along the 3% (or more) credit card processing fees to their customers. Previously many credit card companies would penalize vendors for doing this (and possibly cut them off). This fee was considered "the cost of doing business".

      Some vendors prior to the pandemic would provide cash discounts on large orders because they could circumvent these fees.

      Does this affect (harm) inflation? Is it a form of Gresham's law at play here? What effect does this have on credit card companies? Are they so integral to the system that it doesn't affect them, but instead the customers using their legal tender?

  22. Feb 2023
    1. Remember that life in a Zettelkasten is supposed to be fun. It is a joyful experience to work with it when it works back with you. Life in Zettelkasten is more like dance than a factory.

      I've always disliked the idea of "work" involved in "making" notes and "processing" them. Framing zettelkasten and knowledge creation in terms of capitalism is a painful mistake.

      the quote is from https://blay.se/2015-06-21-living-with-a-zettelkasten.html

    1. Deutsch’s index was created out of an almost algorith-mic processing of historical sources in the pursuit of a totalized and perfect history of theJews; it presented, on one hand, the individualized facts, but together also constitutedwhat we might term a ‘history without presentation’, which merely held the ‘facts’themselves without any attempt to synthesize them (cf. Saxer, 2014: 225-32).

      Not sure that I agree with the framing of "algorithmic processing" here as it was done manually by a person pulling out facts. But it does bring out the idea of where collecting ends and synthesis of a broader thesis out of one's collection begins. Where does historical method end? What was the purpose of the collection? Teaching, writing, learning, all, none?

    1. rank is not an assessment of who has thebest intrinsic properties, but rather a useful consensus view thatprovides rules for how to behave toward others.

      Rank (social or otherwise) can be a signal for predictability from the perspective of consensus views for how to behave towards others with respect to the abilities or values being measured.


      Ranking people for some sort of technical ability may be a better objective measure rather than ranking people on social status which is far less objective from a humanist perspective. In employment situations, individuals are more likely to rely on social and cultural biases and racist tendencies rather than on objective measures with respect to the job at hand. How can we better objectify the actual underlying values over and above the more subjective ones.

  23. Jan 2023
    1. a common technique in natural language processing is to operationalize certain semantic concepts (e.g., "synonym") in terms of syntactic structure (two words that tend to occur nearby in a sentence are more likely to be synonyms, etc). This is what word2vec does.

      Can I use some of these sorts of methods with respect to corpus linguistics over time to better identified calcified words or archaic phrases that stick with the language, but are heavily limited to narrower(ing) contexts?

    1. Fried-berg Judeo-Arabic Project, accessible at http://fjms.genizah.org. This projectmaintains a digital corpus of Judeo-Arabic texts that can be searched and an-alyzed.

      The Friedberg Judeo-Arabic Project contains a large corpus of Judeo-Arabic text which can be manually searched to help improve translations of texts, but it might also be profitably mined using information theoretic and corpus linguistic methods to provide larger group textual translations and suggestions at a grander scale.

  24. Dec 2022
  25. Nov 2022
    1. partnerships, networking, and revenue generation such as donations, memberships, pay what you want, and crowdfunding

      I have thought long about the same issue and beyond. The triple (wiki, Hypothesis, donations) could be a working way to search for OER, form a social group processing them, and optionally support the creators.

      I imagine that as follows: a person wants to learn about X. They can head to the wiki site about X and look into its Hypothesis annotations, where relevant OER with their preferred donation method can be linked. Also, study groups interested in the respective resource or topic can list virtual or live meetups there. The date of the meetups could be listed in a format that Hypothesis could search and display on a calendar.

      Wiki is integral as it categorizes knowledge, is comprehensive, and strives to address biases. Hypothesis stitches websites together for the benefit of the site owners and the collective wisdom that emerges from the discussions. Donations support the creators so they can dedicate their time to creating high-quality resources.

      Main inspirations:

      Deschooling Society - Learning Webs

      Building the Global Knowledge Graph

      Schoolhouse calendar

    1. first we're looking for the "main" object. The word "main" is used in lots of places in Ruby, so that will be hard to track down. How else can we search?Luckily, we know that if you print out that object, it says "main". Which means we should be able to find the string "main", quotes and all, in C.
    1. Robert Amsler is a retired computational lexicology, computational linguist, information scientist. His P.D. was from UT-Austin in 1980. His primary work was in the area of understanding how machine-readable dictionaries could be used to create a taxonomy of dictionary word senses (which served as the motivation for the creation of WordNet) and in understanding how lexicon can be extracted from text corpora. He also invented a new technique in citation analysis that bears his name. His work is mentioned in Wikipedia articles on Machine-Readable dictionary, Computational lexicology, Bibliographic coupling, and Text mining. He currently lives in Vienna, VA and reads email at robert.amsler at utexas. edu. He is currenly interested in chronological studies of vocabulary, esp. computer terms.

      https://www.researchgate.net/profile/Robert-Amsler

      Apparently follow my blog. :)

      Makes me wonder how we might better process and semantically parse peoples' personal notes, particularly when they're atomic and cross-linked?

  26. Oct 2022
    1. elaboration n. 1. the process of interpreting or embellishing information to be remembered or of relating it to other material already known and in memory. The levels-of-processing model of memory holds that the level of elaboration applied to information as it is processed affects both the length of time that it can be retained in memory and the ease with which it can be retrieved.

      https://dictionary.apa.org/elaboration

    1. https://www.explainpaper.com/

      Another in a growing line of research tools for processing and making sense of research literature including Research Rabbit, Connected Papers, Semantic Scholar, etc.

      Functionality includes the ability to highlight sections of research papers with natural language processing to explain what those sections mean. There's also a "chat" that allows you to ask questions about the paper which will attempt to return reasonable answers, which is an artificial intelligence sort of means of having an artificial "conversation with the text".

      cc: @dwhly @remikalir @jeremydean

  27. Sep 2022
    1. maintenance rehearsal repeating items over and over to maintain them in short-term memory, as in repeating a telephone number until it has been dialed (see rehearsal). According to the levels-of-processing model of memory, maintenance rehearsal does not effectively promote long-term retention because it involves little elaboration of the information to be remembered. Also called rote rehearsal. See also phonological loop.

      The practice of repeating items as a means of attempting to place them into short-term memory is called maintenance rehearsal. Examples of this practice include repeating a new acquaintance's name or perhaps their phone number multiple times as a means of helping to remember it either for the short term or potentially the long term.

      Research on the levels-of processing model of memory indicates that maintenance rehearsal is not as effective at promoting long term memory as methods like elaborative rehearsal.

  28. Aug 2022
  29. Jul 2022
  30. Jun 2022
    1. We are the leading independent Open Access publisher in the Humanities and Social Sciences in the UK: a not-for-profit Social Enterprise run by scholars who are committed to making high-quality research freely available to readers around the world. All our books are available to read online and download for free, with no Book Processing Charges (BPCs) for authors. We publish monographs and textbooks in all areas, offering the academic excellence of a traditional press combined with the speed, convenience and accessibility of digital publishing. We also publish bespoke Series for Universities and Research Centers and invite libraries to support Open Access publishing by joining our Membership Programme.

      https://www.openbookpublishers.com/

    1. The absence of Quick Note on the iPhone is a strange, glaring omission that’s baffling to me. I do research on every device, including the iPhone. In fact, I’d argue that the iPhone is the most important place to include Quick Note. That’s because, despite the ample screen of my iPhone 12 Pro Max, it’s still not the best place to read, making saving items for later with Quick Note more valuable there. However, my iPhone is still where I run across links and other material I want to save daily. I’d love to be able to drop links and blockquotes into Quick Note from my iPhone, so I could revisit the material later from the more comfortable reading environment of my iPad or Mac. Not having Quick Note on the iPhone is a significant blow to the feature’s utility.

      Considering how I've been publicly speaking and behaving (melodramatically, that is) - as someone who has returned to using my iPhone as my primary working device - this sort of oversight is precisely what I expected, actually, What I did not expect of Apple was to respond as early as the next numeric release to this omission.

      Running this very first build of iOS 16, I can indeed that Apple has thought of at least one original context for Quick Note creation, but obviously, it's quite hard to say at this point.

      Anywho/how, here's what it looks like at the moment.

      Quick Note implemented on iPhone as of iOS 16's very first available dev beta

  31. May 2022
    1. .Adopting the habit of knowledge capture has immediate benefitsfor our mental health and peace of mind. We can let go of the fearthat our memory will fail us at a crucial moment. Instead of jumpingat every new headline and notification, we can choose to consumeinformation that adds value to our lives and consciously let go of therest.

      Immediate knowledge capture by highlighting, annotating, or other means when taking notes can help to decrease cognitive load. This is similar to other productivity methods like quick logging within a bullet journal system, writing morning pages, or Getting Things Done (GTD). By putting everything down in one place, you can free your mind of the constant need to remember dozens of things. This frees up your working memory to decrease stress as you know you've captured the basic idea for future filtering, sorting, and work at a later date.

  32. Mar 2022
  33. Jan 2022
    1. Fernandez-Castaneda, A., Lu, P., Geraghty, A. C., Song, E., Lee, M.-H., Wood, J., Yalcin, B., Taylor, K. R., Dutton, S., Acosta-Alvarez, L., Ni, L., Contreras-Esquivel, D., Gehlhausen, J. R., Klein, J., Lucas, C., Mao, T., Silva, J., Pena-Hernandez, M., Tabachnikova, A., … Monje, M. (2022). Mild respiratory SARS-CoV-2 infection can cause multi-lineage cellular dysregulation and myelin loss in the brain (p. 2022.01.07.475453). https://doi.org/10.1101/2022.01.07.475453

    1. Most developers are familiar with MySQL and PostgreSQL. They are great RDBMS and can be used to run analytical queries with some limitations. It’s just that most relational databases are not really designed to run queries on tens of millions of rows. However, there are databases specially optimized for this scenario - column-oriented DBMS. One good example is of such a database is ClickHouse.

      How to use Relational Databases to process logs

  34. Dec 2021
    1. Catala, a programming language developed by Protzenko's graduate student Denis Merigoux, who is working at the National Institute for Research in Digital Science and Technology (INRIA) in Paris, France. It is not often lawyers and programmers find themselves working together, but Catala was designed to capture and execute legal algorithms and to be understood by lawyers and programmers alike in a language "that lets you follow the very specific legal train of thought," Protzenko says.

      A domain-specific language for encoding legal interpretations.

    1. One more thing ought to be explained in advance: why the card index is indeed a paper machine. As we will see, card indexes not only possess all the basic logical elements of the universal discrete machine — they also fi t a strict understanding of theoretical kinematics . The possibility of rear-ranging its elements makes the card index a machine: if changing the position of a slip of paper and subsequently introducing it in another place means shifting other index cards, this process can be described as a chained mechanism. This “ starts moving when force is exerted on one of its movable parts, thus changing its position. What follows is mechanical work taking place under particular conditions. This is what we call a machine . ” 11 The force taking effect is the user ’ s hand. A book lacks this property of free motion, and owing to its rigid form it is not a paper machine.

      The mechanical work of moving an index card from one position to another (and potentially changing or modifying links to it in the process) allows us to call card catalogues paper machines. This property is not shared by information stored in codices or scrolls and thus we do not call books paper machines.

  35. Nov 2021
  36. Oct 2021
  37. Sep 2021
  38. Jul 2021
  39. Jun 2021
  40. May 2021
  41. Apr 2021
    1. This post articulates a lot of what I've been thinking about for the past 18 months or so, but it adds the additional concept of community integration.

      Interestingly, this aligns with the early, tentative ideas around what the future of In Beta might look like as a learning community, rather than a repository of content.

  42. Mar 2021
  43. Jan 2021
  44. Nov 2020
  45. Oct 2020
  46. Sep 2020
  47. Aug 2020
  48. Jul 2020
    1. As mentioned earlier in these guidelines, it is very important that controllers assess the purposes forwhich data is actually processed and the lawful grounds on which it is based prior to collecting thedata. Often companies need personal data for several purposes, and the processing is based on morethan one lawful basis, e.g. customer data may be based on contract and consent. Hence, a withdrawalof consent does not mean a controller must erase data that are processed for a purpose that is basedon the performance of the contract with the data subject. Controllers should therefore be clear fromthe outset about which purpose applies to each element of data and which lawful basis is being reliedupon.
    2. In cases where the data subject withdraws his/her consent and the controller wishes to continue toprocess the personal data on another lawful basis, they cannot silently migrate from consent (which iswithdrawn) to this other lawful basis. Any change in the lawful basis for processing must be notified toa data subject in accordance with the information requirements in Articles 13 and 14 and under thegeneral principle of transparency.
    1. Some vendors may relay on legitimate interest instead of consent for the processing of personal data. The User Interface specifies if a specific vendor is relating on legitimate interest as legal basis, meaning that that vendor will process user’s data for the declared purposes without asking for their consent. The presence of vendors relying on legitimate interest is the reason why within the user interface, even if a user has switched on one specific purpose, not all vendors processing data for that purpose will be displayed as switched on. In fact, those vendors processing data for that specific purpose, relying only on legitimate interest will be displayed as switched off.
  49. Jun 2020
  50. May 2020
    1. Sure, anti-spam measures such as a CAPTCHA would certainly fall under "legitimate interests". But would targeting cookies? The gotcha with reCAPTCHA is that this legitimate-interest, quite-necessary-in-today's-world feature is inextricably bundled with unwanted and unrelated Google targeting (cookiepedia.co.uk/cookies/NID) cookies (_ga, _gid for v2; NID for v3).
    1. there’s no need to send consent request emails — provided that this basis of processing was stated in your privacy policy and that users had easy access to the notice prior to you processing their data. If this information was not available to users at the time, but one of these legal bases can currently legitimately apply to your situation, then your best bet would be to ensure that your current privacy notice meets requirements, so that you can continue to process your user data in a legally compliant way.
    1. With that guidance in mind, and from a practical standpoint, consider keeping records of the following: The name or other identifier of the data subject that consented; The dated document, a timestamp, or note of when an oral consent was made; The version of the consent request and privacy policy existing at the time of the consent; and, The document or data capture form by which the data subject submitted his or her data.
    1. “Until CR 1.0 there was no effective privacy standard or requirement for recording consent in a common format and providing people with a receipt they can reuse for data rights.  Individuals could not track their consents or monitor how their information was processed or know who to hold accountable in the event of a breach of their privacy,” said Colin Wallis, executive director, Kantara Initiative.  “CR 1.0 changes the game.  A consent receipt promises to put the power back into the hands of the individual and, together with its supporting API — the consent receipt generator — is an innovative mechanism for businesses to comply with upcoming GDPR requirements.  For the first time individuals and organizations will be able to maintain and manage permissions for personal data.”
    1. It’s useful to remember that under GDPR regulations consent is not the ONLY reason that an organization can process user data; it is only one of the “Lawful Bases”, therefore companies can apply other lawful (within the scope of GDPR) bases for data processing activity. However, there will always be data processing activities where consent is the only or best option.
    1. If you’re a controller based outside of the EU, you’re transferring personal data outside of the EU each time you collect data of users based within the EU. Please make sure you do so according to one of the legal bases for transfer.

      Here they equate collection of personal data with transfer of personal data. But this is not very intuitive: I usually think of collection of data and transfer of data as rather different activities. It would be if we collected the data on a server in EU and then transferred all that data (via some internal process) to a server in US.

      But I guess when you collect the data over the Internet from a user in a different country, the data is technically being transferred directly to your server in the US. But who is doing the transfer? I would argue that it is not me who is transferring it; it is the user who transmitted/sent the data to my app. I'm collecting it from them, but not transferring it. Collecting seems like more of a passive activity, while transfer seems like a more active activity (maybe not if it's all automated).

      So if these terms are equivalent, then they should replace all instances of "transfer" with "collect". That would make it much clearer and harder to mistakenly assume this doesn't apply to oneself. Or if there is a nuanced difference between the two activities, then the differences should be explained, such as examples of when collection may occur without transfer occurring.

    1. Full and extensive records of processing are expressly required in cases where your data processing activities are not occasional, where they could result in a risk to the rights and freedoms of others, where they involve the handling of “special categories of data” or where your organization has more than 250 employees — this effectively covers almost all data controllers and processors.
    1. it buys, receives, sells, or shares the personal information of 50,000 or more consumers annually for the business’ commercial purposes. Since IP addresses fall under what is considered personal data — and “commercial purposes” simply means to advance commercial or economic interests — it is likely that any website with at least 50k unique visits per year from California falls within this scope.
    1. You must disclose how the add-on collects, uses, stores and shares user data in the privacy policy field on AMO. Mozilla expects that the add-on limits data collection whenever possible, in keeping with Mozilla’s Lean Data Practices and Mozilla’s Data Privacy Principles, and uses the data only for the purpose for which it was originally collected.
  51. Apr 2020
    1. organizations may require consent from individuals where the processing of personal data is likely to result in a risk or high risk to the rights and freedoms of individuals or in the case of automated individual decision-making and profiling. Formal consent could as well be justified where the processing requires sharing of personal data with third parties, international data transfers, or where the organization processes special categories of personal data or personal data from minors.
    1. In geochemistry, we know that around US$7,000,000 each year is spent on open access to journals [9], with virtually none of this being reinvested into the community itself or the community being reimbursed. Given the immense value of preprints, reinvesting this value into more sustainable community-led non-profit ventures, such as EarthArXiv, is of great potential.
      • Biaya (dalam bentuk APC) yang dikeluarkan untuk menerbitkan makalah sangat tinggi. Biaya tersebut adalah di luar (on top) dari biaya riset yang telah dikeluarkan oleh peneliti atau lembaga pemberi dana riset (funder).
      • Biaya publikasi merupakan proporsi anggaran untuk sebuah dokumen yang berada di bagian akhir dari siklus riset, bukan anggaran inti.
      • Akan lebih baik kalau anggaran publikasi tersebut, sebagian besar atau seluruhnya dialirkan untuk membiayai kegiatan inti, yaitu riset.
      • Referensi: MDPI APC, NCBI, Tabel, King 2007, Calaos, 2011
    1. The data is stored in log files to ensure the functionality of the website. In addition, the data serves us to optimize the website and to ensure the security of our information technology systems. An evaluation of the data for marketing purposes does not take place in this context. The legal basis for the temporary storage of the data and the log files is Art. 6 para. 1 lit. f GDPR. Our legitimate interests lie in the above-mentioned purposes.