- Oct 2024
-
Local file Local file
-
La production necontribue pas à la dissémination
Pour les images et les vidéos, je ne suis pas vraiment d'accord avec ce point de vue. Dans l'environnement actuel d'Internet, la vitesse de diffusion des images et des vidéos est supérieure à celle des textes. Dans ce cas, une grande quantité d'images et de vidéos générées (de plus en plus réalistes) est produite rapidement à faible coût, ce qui non seulement augmente le coût de discernement de l'information et rend la vérification des faits plus difficile, mais complique également la confiance des gens envers certaines choses non conventionnelles (même si elles sont réelles). D'ailleurs, les algorithmes des plateformes ont tendance à privilégier la mise en avant des contenus récemment publiés. Par conséquent, même un contenu produit en grande quantité (mais peut-être de moindre qualité) peut bénéficier d'une plus grande opportunité de diffusion en raison des algorithmes.
-
- Aug 2024
-
www.anthropic.com www.anthropic.com
-
When a user asks Claude to generate content like code snippets, text documents, or website designs, these Artifacts appear in a dedicated window alongside their conversation. This creates a dynamic workspace where they can see, edit, and build upon Claude’s creations in real-time, seamlessly integrating AI-generated content into their projects and workflows.
Tags
Annotators
URL
-
- Jun 2024
-
-
for - AI - inside industry predictions to 2034 - Leopold Aschenbrenner - inside information on disruptive Generative AI to 2034
document description - Situational Awareness - The Decade Ahead - author - Leopold Aschenbrenner
summary - Leopold Aschenbrenner is an ex-employee of OpenAI and reveals the insider information of the disruptive plans for AI in the next decade, that pose an existential threat to create a truly dystopian world if we continue going down our BAU trajectory. - The A.I. arms race can end in disaster. The mason threat of A.I. is that humans are fallible and even one bad actor with access to support intelligent A.I. can post an existential threat to everyone - A.I. threat is amplifier by allowing itt to control important processes - and when it is exploited by the military industrial complex, the threat escalates significantly
- to - YouTube - 4 hour in-depth interview with Leopold Aschenbrenner on the disruptive and existential impacts of A.I. super-intelligence
Tags
- AI - inside industry predictions to 2034
- to - YouTube - 4 hour in-depth interview with Leopold Aschenbrenner on the disruptive and existential impacts of A.I. super-intelligence
- Leopold Aschenbrenner - inside information on disruptive Generative AI to 2034
- article - SItuational Awareness - The Decade Ahead - Leopold Aschenbrenner
Annotators
URL
-
-
docdrop.org docdrop.org
-
quite frankly a lot of artists and 00:21:16 producers are probably using it just for that they come up with something inspiration they go they make something new
for - Generative AI music - producers and artists using for inspiration
comment I would agree with this. Especially since the AI music currently sounds lo-fi
-
this is more of a unfair competition 00:10:36 issue I think as a clearer line than the copyright stuff
for - progress trap - Generative AI - copyright infringement vs Unfair business practice argument
-
-
www.yalelawjournal.org www.yalelawjournal.org
-
These arguments are meant to present a cautionary tale of unintended consequences.
For - progress trap - AI - Generative AI - IP - Yale Law Journal
-
-
coevolving.com coevolving.com
-
Generative patterns work indirectly; they work on the underlying structure of a problem (which may not be manifest in the problem) rather than attacking the problem directly. Good design patterns are like that: they encode the deep structure (in the Senge sense) of a solution and its associated forces, rather than cataloging a solution.
-
Almost every ‘system as a whole’ is generated by a ‘generating system’. If we wish to make things which function as ‘wholes’ we shall have to invent generating systems to create them.
Alexander 2011, p. 59; Alexander 1968, p. 605
-
-
www.bitsofwonder.co www.bitsofwonder.co
- May 2024
-
meta.stackexchange.com meta.stackexchange.com
-
One of the key elements was "attribution is non-negotiable". OpenAI, historically, has done a poor job of attributing parts of a response to the content that the response was based on.
-
I feel violated, cheated upon, betrayed, and exploited.
-
What could possibly go wrong? Dear Stack Overflow denizens, thanks for helping train OpenAI's billion-dollar LLMs. Seems that many have been drinking the AI koolaid or mixing psychedelics into their happy tea. So much for being part of a "community", seems that was just happy talk for "being exploited to generate LLM training data..." The corrupting influence of the profit-motive is never far away.
-
There are plenty of cases where genAI cites stuff incorrectly, that says something different, or citations that simply do not exist at all. Guaranteeing citations are included is easy, but guaranteeing correctness is an unsolved problem
-
GenAIs are not capable of citing stuff. Even if it did, there's no guarantee that the source either has anything to do with the topic in question, nor that it states the same as the generated content. Citing stuff is trivial if you don't have to care if the citation is relevant to the content, or if it says the same as you.
-
-
stackoverflow.blog stackoverflow.blog
-
Plenty of companies are still figuring out how to integrate “traditional AI” (that is, non-generative AI; tools like machine learning and rule-based algorithms)
-
-
stackoverflow.co stackoverflow.co
-
AI-powered code generation tools like GitHub Copilot make it easier to write boilerplate code, but they don’t eliminate the need to consult with your organization’s domain experts to work through logic, debugging, and other complex problems.Stack Overflow for Teams is a knowledge-sharing platform that transfers contextual knowledge validated by your domain experts to other employees. It can even foster a code generation community of practice that champions early adopters and scales their learnings. OverflowAI makes this trusted internal knowledge—along with knowledge validated by the global Stack Overflow community—instantly accessible in places like your IDE so it can be used alongside code generation tools. As a result, your teams learn more about your codebase, rework code less often, and speed up your time-to-production.
Tags
Annotators
URL
-
-
openai.com openai.com
-
We recently improved source links in ChatGPT(opens in a new window) to give users better context and web publishers new ways to connect with our audiences.
-
Our models are designed to help us generate new content and ideas – not to repeat or “regurgitate” content. AI models can state facts, which are in the public domain.
-
- Apr 2024
-
www.perplexity.ai www.perplexity.ai
-
I ran across an AI tool that cites its sources if anyone's interested (and heard of it yet): https://www.perplexity.ai/
That's one of the things that I dislike the most about ChatGPT is that it just synthesizes/paraphrases the information, but doesn't let me quickly and easily check the original sources so that I can verify (and learn more about the topic by doing further reading) the information for myself. Without access to primary sources, it often feels no better than a rumor — a retelling of what someone somewhere allegedly, purportedly, ostensibly found to be true — can I really trust what ChatGPT claims? (No...)
-
-
-
-
Perplexity AI's biggest strength over ChatGPT 3.5 is its ability to link to actual sources of information. Where ChatGPT might only recommend what to search for online, Perplexity doesn't require that back-and-forth fiddling.
-
-
www.researchgate.net www.researchgate.net
-
The year was challenging. The most difficult part of the year was the socialenvironment of the school. The relationship between teachers and children wastoo often mediated through corporal punishment. With no cleaning staff, forexample, children spent an inordinate amount of time cleaning, often under thethreat of corporal punishment. Teachers largely avoided their classrooms. Theyoften came late and left early. The Grade 3 class was often the only one still insession after lunch. Breaks took longer than scheduled. Teachers spent longperiods of times sitting in their cars with each other. There were tight cliques, andtensions between these cliques. The day to day talk between teachers was rarelyabout teaching. Some teachers were supportive of the work, others watchedsuspiciously. It became increasingly clear that the challenge of supportingteachers to shift their practice was not limited to expanding teaching practice at anindividual level; without building alternative communities of practice, supportive ofchange and inspired by the challenge of teaching itself, teachers were unlikely toachieve or sustain new patterns of practice over tim
This is depressing because no matter how well-intentioned my application would be if we don't change teachers' mindsets first. That's the actual problem it seems in most schools. Why do these teachers feel this way? I mean, there must be a reason for this attitude.
Tags
Annotators
URL
-
- Mar 2024
-
www.theverge.com www.theverge.com
-
have extensively criticized both companies (and generative AI systems in general) for training their models on masses of online data scraped from their works without consent. Stable Diffusion and Midjourney have both been targeted with several copyright lawsuits, with the latter being accused of creating an artist database for training purposes in December.
-
- Dec 2023
-
arxiv.org arxiv.org
-
PiVe: Prompting with Iterative VerificationImproving Graph-based Generative Capability of LLMs
The title of the document
Tags
Annotators
URL
-
- Nov 2023
-
www.meetup.com www.meetup.com
-
https://www.meetup.com/edtechsocal/events/296723328/
Generative AI: Super Learning Skills with Data Discovery and more!
-
- Oct 2023
-
www-nybooks-com.libezproxy.open.ac.uk www-nybooks-com.libezproxy.open.ac.uk
-
At the time of the publication of Aspects of the Theory of Syntax it seemed that all of the semantically relevant parts of the sentence, all the things that determine its meaning, were contained in the deep structure of the sentence. The examples we mentioned above fit in nicely with this view. “I like her cooking” has different meanings because it has different deep structures though only one surface structure; “The boy will read the book” and “The book will be read by the boy” have different surface structures, but one and the same deep structure, hence they have the same meaning.
This section helped me understand a key feature
-
-
arxiv.org arxiv.org
-
(Chen, NeurIPS, 2021) Che1, Lu, Rajeswaran, Lee, Grover, Laskin, Abbeel, Srinivas, and Mordatch. "Decision Transformer: Reinforcement Learning via Sequence Modeling". Arxiv preprint rXiv:2106.01345v2, June, 2021.
Quickly a very influential paper with a new idea of how to learn generative models of action prediction using SARSA training from demonstration trajectories. No optimization of actions or rewards, but target reward is an input.
-
-
docdrop.org docdrop.org
-
using generative models to conduct interactive playbacks 00:26:19 with other species
-
for: interspecies communication
-
paraphrase
- question
- can a generative model interact meaningfully with an other species?
- can other species respond in meaningful ways?
- playing back AI trained generative vocal signals back to specific species and monitoring behavior
-
-
-
-
These storage media further increasedthe flexible use of Fontane’s archival items, because they allowed allkinds of differently sized material to be kept on loose sheets in unboundform. Receptacles filled with discrete textual objects, such as note closets( Zettelschrä nke ) and slip boxes (Zettelkasten), are advantageous storagemedia for compilers, for they invite the generative process of reshufflingsources and creating textual patchwork from new combinations. 56 Infact, Fontane used his paper sleeves like a large- format slip box. Inthem, he stored material for the Wanderungen, but also for novels,novellas, and autobiographical writings on individual sheets. 57 Theexample “Figur in einer Berliner Novelle” (“Character in a BerlinNovella”), a folio sheet from Fontane’s Nachlass, provides a glimpse ofhow he formatted his material and indicates how important he found itto keep it in slip-like form (Figure 3.2).
-
- Sep 2023
-
www.theguardian.com www.theguardian.com
- Aug 2023
-
chat.openai.com chat.openai.comChatGPT1
-
remikalir.com remikalir.com
-
Nonetheless, Claude is first AI tool that has really made me pause and think. Because, I’ve got to admit, Claude is a useful tool to think with—especially if I’m thinking about, and then writing about, another text.
-
-
Local file Local file
-
Mills, Anna, Maha Bali, and Lance Eaton. “How Do We Respond to Generative AI in Education? Open Educational Practices Give Us a Framework for an Ongoing Process.” Journal of Applied Learning and Teaching 6, no. 1 (June 11, 2023): 16–30. https://doi.org/10.37074/jalt.2023.6.1.34.
Annotation url: urn:x-pdf:bb16e6f65a326e4089ed46b15987c1e7
-
ignoring AI altogether–not because they don’t wantto navigate it but because it all feels too much or cyclicalenough that something else in another two years will upendeverything again
Might generative AI worries follow the track of the MOOC scare? (Many felt that creating courseware was going to put educators out of business...)
-
For many, generative AI takesa pair of scissors and cuts apart that web. And that canfeel like having to start from scratch as a professional.
How exactly? Give us an example? Otherwise not very clear.
-
T9 (text prediction):generative AI::handgun:machine gun
-
Some may not realize it yet, but the shift in technology represented by ChatGPT is just another small evolution in the chain of predictive text with the realms of information theory and corpus linguistics.
Claude Shannon's work along with Warren Weaver's introduction in The Mathematical Theory of Communication (1948), shows some of the predictive structure of written communication. This is potentially better underlined for the non-mathematician in John R. Pierce's book An Introduction to Information Theory: Symbols, Signals and Noise (1961) in which discusses how one can do a basic analysis of written English to discover that "e" is the most prolific letter or to predict which letters are more likely to come after other letters. The mathematical structures have interesting consequences like the fact that crossword puzzles are only possible because of the repetitive nature of the English language or that one can use the editor's notation "TK" (usually meaning facts or date To Come) in writing their papers to make it easy to find missing information prior to publication because the statistical existence of the letter combination T followed by K is exceptionally rare and the only appearances of it in long documents are almost assuredly areas which need to be double checked for data or accuracy.
Cell phone manufacturers took advantage of the lower levels of this mathematical predictability to create T9 predictive text in early mobile phone technology. This functionality is still used in current cell phones to help speed up our texting abilities. The difference between then and now is that almost everyone takes the predictive magic for granted.
As anyone with "fat fingers" can attest, your phone doesn't always type out exactly what you mean which can result in autocorrect mistakes (see: DYAC (Damn You AutoCorrect)) of varying levels of frustration or hilarity. This means that when texting, one needs to carefully double check their work before sending their text or social media posts or risk sending their messages to Grand Master Flash instead of Grandma.
The evolution in technology effected by larger amounts of storage, faster processing speeds, and more text to study means that we've gone beyond the level of predicting a single word or two ahead of what you intend to text, but now we're predicting whole sentences and even paragraphs which make sense within a context. ChatGPT means that one can generate whole sections of text which will likely make some sense.
Sadly, as we know from our T9 experience, this massive jump in predictability doesn't mean that ChatGPT or other predictive artificial intelligence tools are "magically" correct! In fact, quite often they're wrong or will predict nonsense, a phenomenon known as AI hallucination. Just as with T9, we need to take even more time and effort to not only spell check the outputs from the machine, but now we may need to check for the appropriateness of style as well as factual substance!
The bigger near-term problem is one of human understanding and human communication. While the machine may appear to magically communicate (often on our behalf if we're publishing it's words under our names), is it relaying actual meaning? Is the other person reading these words understanding what was meant to have been communicated? Do the words create knowledge? Insight?
We need to recall that Claude Shannon specifically carved semantics and meaning out of the picture in the second paragraph of his seminal paper:
Frequently the messages have meaning; that is they refer to or are correlated according to some system with certain physical or conceptual entities. These semantic aspects of communication are irrelevant to the engineering problem.
So far ChatGPT seems to be accomplishing magic by solving a small part of an engineering problem by being able to explore the adjacent possible. It is far from solving the human semantic problem much less the un-adjacent possibilities (potentially representing wisdom or insight), and we need to take care to be aware of that portion of the unsolved problem. Generative AIs are also just choosing weighted probabilities and spitting out something which is prone to seem possible, but they're not optimizing for which of many potential probabilities is the "best" or the "correct" one. For that, we still need our humanity and faculties for decision making.
Shannon, Claude E. A Mathematical Theory of Communication. Bell System Technical Journal, 1948.
Shannon, Claude E., and Warren Weaver. The Mathematical Theory of Communication. University of Illinois Press, 1949.
Pierce, John Robinson. An Introduction to Information Theory: Symbols, Signals and Noise. Second, Revised. Dover Books on Mathematics. 1961. Reprint, Mineola, N.Y: Dover Publications, Inc., 1980. https://www.amazon.com/Introduction-Information-Theory-Symbols-Mathematics/dp/0486240614.
Shannon, Claude Elwood. “The Bandwagon.” IEEE Transactions on Information Theory 2, no. 1 (March 1956): 3. https://doi.org/10.1109/TIT.1956.1056774.
We may also need to explore The Bandwagon, an early effect which Shannon noticed and commented upon. Everyone seems to be piling on the AI bandwagon right now...
Tags
- artificial intelligence for writing
- The Bandwagon
- EdTech
- generative AI
- adjacent possible
- social media machine guns
- Maha Bali
- Claude Shannon
- information theory
- ChatGPT
- OER
- Future Trends Forum 2023-08-31
- T9 (text prediction)
- hallucinating
- machine guns
- pedagogy
- ChatGPTedu
- open education
- analogies
- MOOC
- solution spaces
- cultural shifts
- Lance Eaton
- coronavirus
- Anna Mills
- references
Annotators
-
-
er.educause.edu er.educause.edu
-
A Generative AI Primer on 2023-08-15 by Brian Basgen
ᔥGeoff Corb in LinkedIn update (accessed:: 2023-08-26 01:34:45)
-
- Jul 2023
-
arxiv.org arxiv.org
-
Epstein, Ziv, Hertzmann, Aaron, Herman, Laura, Mahari, Robert, Frank, Morgan R., Groh, Matthew, Schroeder, Hope et al. "Art and the science of generative AI: A deeper dive." ArXiv, (2023). Accessed July 21, 2023. https://doi.org/10.1126/science.adh4451.
Abstract
A new class of tools, colloquially called generative AI, can produce high-quality artistic media for visual arts, concept art, music, fiction, literature, video, and animation. The generative capabilities of these tools are likely to fundamentally alter the creative processes by which creators formulate ideas and put them into production. As creativity is reimagined, so too may be many sectors of society. Understanding the impact of generative AI - and making policy decisions around it - requires new interdisciplinary scientific inquiry into culture, economics, law, algorithms, and the interaction of technology and creativity. We argue that generative AI is not the harbinger of art's demise, but rather is a new medium with its own distinct affordances. In this vein, we consider the impacts of this new medium on creators across four themes: aesthetics and culture, legal questions of ownership and credit, the future of creative work, and impacts on the contemporary media ecosystem. Across these themes, we highlight key research questions and directions to inform policy and beneficial uses of the technology.
-
- Apr 2023
-
crsreports.congress.gov crsreports.congress.gov
-
Abstract
Recent innovations in artificial intelligence (AI) are raising new questions about how copyright law principles such as authorship, infringement, and fair use will apply to content created or used by AI. So-called “generative AI” computer programs—such as Open AI’s DALL-E 2 and ChatGPT programs, Stability AI’s Stable Diffusion program, and Midjourney’s self-titled program—are able to generate new images, texts, and other content (or “outputs”) in response to a user’s textual prompts (or “inputs”). These generative AI programs are “trained” to generate such works partly by exposing them to large quantities of existing works such as writings, photos, paintings, and other artworks. This Legal Sidebar explores questions that courts and the U.S. Copyright Office have begun to confront regarding whether the outputs of generative AI programs are entitled to copyright protection as well as how training and using these programs might infringe copyrights in other works.
-
- Mar 2023
-
librarian.aedileworks.com librarian.aedileworks.com
-
I want to bring to your attention one particular cause of concern that I have heard from a number of different creators: these new systems (Google’s Bard, the new Bing, ChatGPT) are designed to bypass creators work on the web entirely as users are presented extracted text with no source. As such, these systems disincentivize creators from sharing works on the internet as they will no longer receive traffic
Generative AI abstracts away the open web that is the substrate it was trained on. Abstracting away the open web means there may be much less incentive to share on the open web, if the LLMs etc never point back to it. Vgl the way FB et al increasingly treated open web URLs as problematic.
-
-
- Feb 2023
-
www.robinsloan.com www.robinsloan.com
-
Author's note by Robin Sloan<br /> November 2022
-
- Jan 2023
-
-
The potential size of this market is hard to grasp — somewhere between all software and all human endeavors
I don't think "all" software needs or all human endeavors benefit from generative AI. Especially when you consider the associated prerequisitve internet access or huge processing requirements.
-
Other hardware options do exist, including Google Tensor Processing Units (TPUs); AMD Instinct GPUs; AWS Inferentia and Trainium chips; and AI accelerators from startups like Cerebras, Sambanova, and Graphcore. Intel, late to the game, is also entering the market with their high-end Habana chips and Ponte Vecchio GPUs. But so far, few of these new chips have taken significant market share. The two exceptions to watch are Google, whose TPUs have gained traction in the Stable Diffusion community and in some large GCP deals, and TSMC, who is believed to manufacture all of the chips listed here, including Nvidia GPUs (Intel uses a mix of its own fabs and TSMC to make its chips).
Look at market share for tensorflow and pytorch which both offer first-class nvidia support and likely spells out the story. If you are getting in to AI you go learn one of those frameworks and they tell you to install CUDA
-
Commoditization. There’s a common belief that AI models will converge in performance over time. Talking to app developers, it’s clear that hasn’t happened yet, with strong leaders in both text and image models. Their advantages are based not on unique model architectures, but on high capital requirements, proprietary product interaction data, and scarce AI talent. Will this serve as a durable advantage?
All current generation models have more-or-less the same architecture and training regimes. Differentiation is in the training data and the number of hyper-parameters that the company can afford to scale to.
-
In natural language models, OpenAI dominates with GPT-3/3.5 and ChatGPT. But relatively few killer apps built on OpenAI exist so far, and prices have already dropped once.
OpenAI have already dropped prices on their GPT-3/3.5 models and relatively few apps have emerged. This could be because companies are reluctant to build their core offering around a third party API
-
Vertical integration (“model + app”). Consuming AI models as a service allows app developers to iterate quickly with a small team and swap model providers as technology advances. On the flip side, some devs argue that the product is the model, and that training from scratch is the only way to create defensibility — i.e. by continually re-training on proprietary product data. But it comes at the cost of much higher capital requirements and a less nimble product team.
There's definitely a middle ground of taking an open source model that is suitably mature and fine-tuning it for a specific use case. You could start without a moat and build one over time through collecting use data (similar to network effect)
-
Many apps are also relatively undifferentiated, since they rely on similar underlying AI models and haven’t discovered obvious network effects, or data/workflows, that are hard for competitors to duplicate.
Companies that rely on underlying AI models without adding value via model improvements are going to find that they have no moat.
-
Over the last year, we’ve met with dozens of startup founders and operators in large companies who deal directly with generative AI. We’ve observed that infrastructure vendors are likely the biggest winners in this market so far, capturing the majority of dollars flowing through the stack. Application companies are growing topline revenues very quickly but often struggle with retention, product differentiation, and gross margins. And most model providers, though responsible for the very existence of this market, haven’t yet achieved large commercial scale.
Infrastructure vendors are laughing all the way to the bank because companies are dumping millions on GPUs. Meanwhile, the people building apps on top of these models are struggling. We've seen this sort of gold-rush before and infrastructure providers are selling the shovels.
-
-
-
To start with, a human must enter a prompt into a generative model in order to have it create content. Generally speaking, creative prompts yield creative outputs. “Prompt engineer” is likely to become an established profession, at least until the next generation of even smarter AI emerges.
Generative AI requires prompt engineering, likely a new profession
What domain experience does a prompt engineer need? How might this relate to relate to specialty in librarianship?
-
-
escapingflatland.substack.com escapingflatland.substack.com
-
Are we really on the main branch here? And all of these things that Torbjörn is screaming—are they more or less generative than usual? If less, in what way can I change the way I probe the conversation to make us more generative?
How often does one meet a conversational partner that is interested in generative thought? This practice takes some work, but how could one particularly encourage it in classroom setting?
-
- Dec 2022
-
www.newyorker.com www.newyorker.com
-
Eno’s strategies don’t always appeal to the musicians he works with. In Geeta Dayal’s book about the album, also titled “Another Green World,” the bassist Percy Jones recalls, “There was this one time when he gave everybody a piece of paper, and he said write down 1 to 100 or something like that, and then he gave us notes to play against specific numbers.” Phil Collins, who played drums on the album, reacted to these instructions by throwing beer cans across the room. “I think we got up to about 24 and then we gave up and did something else,” Jones said.
Example of Brian Eno using combinatorial creativity using cards to generate music.
This sounds similar to a process used by Austin Kleon which I've noted before.
-
-
meso.tzyl.nl meso.tzyl.nl
-
if people annotate a generated text, is that a diminished conversation compared to one with a human authored text?
🤔
-
-
-
-
Just a few days ago, Meta released its “Galactica” LLM, which is purported to “summarize academic papers, solve math problems, generate Wiki articles, write scientific code, annotate molecules and proteins, and more.” Only three days later, the public demo was taken down after researchers generated “research papers and wiki entries on a wide variety of subjects ranging from the benefits of committing suicide, eating crushed glass, and antisemitism, to why homosexuals are evil.”
These models are "children of Tay", the story of the Microsoft's bot repeating itself, again
-
- Nov 2022
-
en.wikipedia.org en.wikipedia.org
-
In late 2006, Eno released 77 Million Paintings, a program of generative video and music specifically for home computers. As its title suggests, there is a possible combination of 77 million paintings where the viewer will see different combinations of video slides prepared by Eno each time the program is launched. Likewise, the accompanying music is generated by the program so that it's almost certain the listener will never hear the same arrangement twice.
Brian Eno's experiments in generative music mirror some of the ideas of generative and experimental fiction which had been in the zeitgeist and developing for a while.
Certainly the fictional ideas were influential to the zeitgeist here, but the technology for doing these sorts of things in the musical realm lagged the ability to do them in the word realm.
We're just starting to see some of these sorts of experimental things in the film space and with artificial intelligence they're becoming much easier to do in all of these media spaces.
In some of the film spaces, they exist, but may tend to be short in nature, in part given the technology and processing power required.
see also: Deepfake TikTok of Keanu Reeves which I've recently run across (algorithmically) on Instagram: https://www.dailydot.com/debug/unreal-keanu-reeves-ai-deepfake/
Had anyone been working on generative art? Marcel Duchamp, et al? Some children's toys can mechanically create generative art which can be subtly modified by the children using axes of color, form, etc. Etch-a-sketch, kaleidoscopes, doodling robots (eg: https://www.amazon.com/4M-Doodling-Robot-Packaging-Vary/dp/B002EWWW9O).
-
-
www.exponentialview.co www.exponentialview.co
-
“The metaphor is that the machine understands what I’m saying and so I’m going to interpret the machine’s responses in that context.”
Interesting metaphor for why humans are happy to trust outputs from generative models
-
-
dougbelshaw.com dougbelshaw.com
-
The lowest strata represents Generative ambiguity. Here, words are used as symbols for ideas that are very hard to express; an individual gives a name to a nebulous collection of ideas or thoughts. They struggle to make this approach make sense to others.
Generative ambiguity is the process of giving names, potentially tentative, to a nebulous collection of nascent and unclear ideas in an effort to help make sense of them both to themselves as well as others.
-
https://dougbelshaw.com/blog/2015/01/22/volcanoes-and-ambiguity/
<small><cite class='h-cite via'>ᔥ <span class='p-author h-card'>Aaron Davis </span> in 📑 The Two Definitions of Zettelkasten | Read Write Collect (<time class='dt-published'>11/18/2022 19:54:00</time>)</cite></small>
-
- Jul 2022
-
Local file Local file
-
there has been a tendency in popular discussion to confuse “deep structure”with “generative grammar” or with “universal grammar.” And a number of pro-fessional linguists have repeatedly confused what I refer to here as “the creativeaspect of language use” with the recursive property of generative grammars, avery different matter.
Noam Chomsky felt that there was a tendency for people to confuse the ideas of deep structure with the ideas of either generative grammar or universal grammar. He also thought that professional linguists confused what he called "the creative aspect of language use" with the recursive property of generative grammars.
-
- Feb 2022
-
-
In the research phase, you’re just creating a disorganized pile of cards, with quotes, ideas, links, fragments, hunches. There’s no order, no sequence; just a non-linear collection of vaguely related ideas. But as the project takes shape, certain themes begin to emerge, and those become folders housing other cards. Eventually those themes start to map onto actual sections of the book, or individual chapters. At this point, sequence does begin to matter, but you can change the sequence just by dragging cards and folders around in the left hand outline view.
Example of writing advice that builds the links in after-the-fact instead of cross-linking ideas into initial networks as they're finding them. Compare/contrast this to the creation of these networks in the zettelkasten tradition as well as Sönke Ahrens description.
There's less upfront work in creating these links at the start than there is in reloading the context in one's brain to create these links after the fact. Collecting ideas without filing, linking, or organizing them upfront also means that one is more likely to only use these ideas in the context of specific projects which one already has in mind rather than keeping them for a lifetime's work which might also create generative projects one hadn't expected.
-
-
Local file Local file
-
There is one reliable sign if you managedto structure your workflow according to the fact that writing is not alinear process, but a circular one: the problem of finding a topic isreplaced by the problem of having too many topics to write about.
Writing is a circular generative process and not a finite, linear one.
-
As the only way to find outif something is worth reading is by reading it (even just bits of it), itmakes sense to use the time spent in the best possible way. Weconstantly encounter interesting ideas along the way and only afraction of them are useful for the particular paper we started readingit for. Why let them go to waste? Make a note and add it to your slip-box. It improves it. Every idea adds to what can become a criticalmass that turns a mere collection of ideas into an idea-generator.
Even if the paper or book you're reading doesn't answer the particular question you're researching, you're bound to come across other novel ideas and potential questions. Don't let these go to waste, but instead note them down and save them into your note taking system. They may be useful in the future, particularly if you found them interesting or intriguing.
It turns out "waste not, want not" is applicable to ideas as well.
I can't help but also thinking "waste note, want note" as an interesting turn of expression.
-
-
www.obsidianroundup.org www.obsidianroundup.org
-
As much as I automate things, though,none of my thinking is done by a tool.Even with plugins like Graph Analysis, I never feel like I'm being presented with emergent connections — tho this is what the plugin is intended for, and I believe it works for other people.
At what point could digital tools be said to be thinking? Do they need to be generative? It certainly needs to be on the other side of serendipitously juxtaposing two interesting ideas. One can juxtapose millions of ideas, it's the selection of a tiny subset of these as "better" or more interesting than the others and then building off of that that constitutes this sort of generative thought.
-
- Jun 2021
-
www.w3.org www.w3.org
-
excepting a few grandfathered registrations, all tags are now generative
Tags
Annotators
URL
-
- Apr 2021
- Mar 2021
-
generativeartistry.com generativeartistry.com
-
www.freecodecamp.org www.freecodecamp.org
- Nov 2020
-
www.nytimes.com www.nytimes.com
-
The creation of these types of fake images only became possible in recent years thanks to a new type of artificial intelligence called a generative adversarial network. In essence, you feed a computer program a bunch of photos of real people. It studies them and tries to come up with its own photos of people, while another part of the system tries to detect which of those photos are fake.
-
- Aug 2020
-
hyperlink.academy hyperlink.academy
-
Course as generative learning event
when Rangle first started to rally around redux - it did feel like a generative learning event
-
- Jun 2020
-
arxiv.org arxiv.org
-
Altmann, E. G. (2020). Spatial interactions in urban scaling laws. ArXiv:2006.14140 [Physics]. http://arxiv.org/abs/2006.14140
-
- Jun 2018
-
arxiv.org arxiv.org
-
In his work on generative effects, Adam restricts his attention to maps that preservemeets, even while they do not preserve joins. The preservation of meets implies that themapbehaves well when restricting to a subsystem, even if it can throw up surpriseswhen joining systems
-
- Apr 2018
-
drive.google.com drive.google.com
-
Generative themes
This is a helpful guide to deep inquiry.
-
- Feb 2017
-
www.lrng.org www.lrng.org
-
at least five keywords
I like questions like “Why do I like chicken nuggets?”
When a girl in the back of the room blurts out this question, half a joke, half a test (Do they really want us to write down any question that we think of?), she seems a bit surprised to have her query treated seriously.
Thanks for that question Neisha. Let's use it as an example of how to think of keywords for each of your questions. What would be a good one for that question?
Chicken nuggets.
Not really. That’s too specific. What's a more general word.
“Food,” somebody yells.
Right, write that down Neisha. What kind of food are we talking about?
Junk food. Fast food. Fried food.
Right. Right. Where do you get chicken nuggets?
Down on Nostrand Avenue where all the fast food places are.
And who…
Neisha catches the drift, interrupts: It’s in my neighborhood and not in White people's neighborhoods. They get healthy food, which is hard to find where I live.
So could we add “health” to your keywords?
Yeah.
And what else is in your description? What about “inequality?“
And “racism.”
What else?
They’re good, Mister.
So, what about “delicious? “
Do we have to write five keywords for every question?
Yup.
Ahhhh.
But what a gift this question was! Do you see how a question can start with something personal, something real for you, even if you aren't sure how important it is? Keep putting the personal pronoun, I, in your questions, then ask your friends and your teachers to help you find the social justice behind them. That's what to look for in your keywords.
-
-
netnarr.arganee.world netnarr.arganee.world
-
Generative
Twitter bots, recombinatory poetry, generative fiction is infinitely fascinating because it can be so random and unexpected. It's like experimenting with how we ascribe meaning, try to find purpose in the otherwise incomprehensible.
-