21 Matching Annotations
  1. Nov 2018
    1. We live in a connected world, and modern software has to navigate this world. So the building blocks for tomorrow's very largest solutions are connected and massively parallel. It's not enough for code to be "strong and silent" any more. Code has to talk to code. Code has to be chatty, sociable, well-connected. Code has to run like the human brain, trillions of individual neurons firing off messages to each other, a massively parallel network with no central control, no single point of failure, yet able to solve immensely difficult problems. And it's no accident that the future of code looks like the human brain, because the endpoints of every network are, at some level, human brains.

      This is an observation that is quite important but frequently missed

  2. Jun 2018
    1. connectionism

      neural nets...

    2. So far, we have dealt with self-reference, but the situation is quite similar with the notion of self-modification. Partial self- modification is easy to achieve; the complete form goes beyond ordinary mathematics and anything we can formulate. Consider, for instance, recursive programs. Every recursive program can be said to modify itself in some sense, since (by the definition of recursiveness) the exact operation carried out at time t depends on the result of the operation at t-1, and so on: therefore, the final "shape" of the transformation is getting defined iteratively, in runtime (a fact somewhat obscured by the usual way in which recursion is written down in high-level programming languages like C). At the same time, as we can expect, to every finite recursive program there belongs an equivalent "straight" program, that uses no recursion at all, and is perfectly well defined in advance, so that it does not change in any respect; it is simply a fixed sequence of a priori given elementary operations.

      So unbounded recursion automatically implies a form of self-reference and self-modification?

  3. Nov 2017
    1. It seems that one-hot encoding isn’t necessary for tree-based algorithms, but you may run into a problem using this on linear models. I personally find ordinal encoding is more than sufficient and it significantly reduces your memory footprint if you have quite a few categorical features.

      Interesting! I suppose it makes sense though.

  4. Aug 2017
    1. This is a very easy paper to follow, but it looks like their methodology is a simple way to improve performance on limited data. I'm curious how well this is reproduced elsewhere.

    2. DEEP learning refers to a network of connected artificialneurons in multiple layers, which can perform featureextraction from observed data and learn the complicatedrelationships among the features of data.

      That's a little constrained of a definition - at least, that's the idea that I have from "Deep Learning" by Goodfellow, Bengio, and Courville. Currently, ANNs are the preferred way (possibly only established way) of doing deep learning, but deep learning is more general than ANNs.

  5. Jul 2017
    1. Kind of a clickbait title - but makes some good points. Personally, though, I considered it common knowledge that storytelling can very effectively incorporate philosophy (and vice versa). Consider Twilight Zone. Consider 12 Angry Men. Consider every movie that has ever presented a moral dilemma and made the viewer think deeply about it.

    1. This is good to know. My page uses Hakyll, and I've used Pandoc many times in the past, and I use org-mode nearly daily... yet somehow I am here fretting that I don't know an easy way to publish from my org-mode documents.

    1. Can I get an RSS feed to follow new annotations on a document? Yes! You can get RSS and Atom feeds by URL, Tag and User.

      Handy! I wonder if I could tie this in with RSS readers to get something akin to Pocket, but not siloed.

    1. The web today is full of disposable speech acts, that are not maintained, enriched, or returned to. Tweets, Facebook posts, contextually dependent blog posts. Consequently entering new conversations feels like sifting through the Nixon tapes.

      This expresses neatly something I've complained about in many other contexts - particularly in trying to piece together coherent information on a project when people haven't taken the time to "garden", so everything is just full of old ephemera at best.

  6. Jan 2017
    1. There is a conceptual and language gap. The sciences of neural networks and probability models do not have a shared language. My goal is to bridge this idea gap and allow for more collaboration and discussion between these fields, and provide a consistent implementation (Github link).

      Compare with https://colah.github.io/posts/2015-09-NN-Types-FP/ and the three narratives (biological, representative, probabilistic)

  7. Dec 2016
    1. Handy! I would like to use this to let Wallabag (which has its own annotation support) send annotations to hypothes.is.

    1. Critics have suggested this way of thinking about the stock market is outdated. In other words, “this time is different.” And even if they admit that comparing equity valuations to net worth has some value they insist that value does not include timing the market.

      Empirically does this fit?

  8. Nov 2016
    1. “10,000 hours” theory of expertise

      Oh, I thought that was Gladwell.

    2. Accept Everything. But Don’t Be Passive

      This seems like a principle that Nassim Nicholas Taken would agree with. He talked at length about stoicism in one of his books - I cannot recall whether Black Swan or Antifragile.

    1. This is a picture of the first HTTP web server in the world. It was Tim Berners-Lee's NeXT computer at CERN. Pasted on the machine is an ominous sticker: "This machine is a server, do not power it down!!". The reason it couldn't be powered down is that web sites on other servers were starting to link to it. Once they linked to it, they then depended on that machine continuing to exist. If the machine was powered down, the links stopped working. If the machine failed or was no longer accessible at the same location, a far worse thing happened: the chain between sites becomes permanently broken, and the ability to access that content is lost forever. That sticker perfectly highlights the biggest problem with HTTP: it erodes.

      This is interesting, since the opening video for https://hypothes.is/ mentions the early web also - in this case, for its annotation features that were removed.

      It seems to me that hypothes.is is even more powerful used on IPFS content identified by hash since that underlying content cannot change.

      Thanks to both services I'm doing exactly this right now!

    2. I think this is exactly what I've wanted - and what a lot of people have wanted - for a long time. It's certainly not the first time I've seen someone call for using hashes for referring to files, but the design and implementation behind this look like they do a lot of things right.

  9. Oct 2016
    1. In the 1970s, software problems were more accessible in the sense that fewer regulations were applied to restrict programmers to access key materials (codes peculiarly). There was a so-called "collaborative hacker" approach, sharing knowledge and improving each other's work, that encouraged programmers to continually redefine the boundaries of the problem (e.g. conducting reverse engineering to deconstruct a software to understand how a code was written). As a result, a wide range of software programming tools (languages, editors, compilers etc.) was created in the 1970s (Ceruzzi, 2003).

      That is a perspective I hadn't considered, despite having read RMS's biography. I had really assumed that "modern" open source led to code being much less encumbered.