-
There have been a number of new AI search engine demos released recently. As I have written about before this is an idea I am especially excited about because it can combine the incredible synthesis abilities of LLMs with verifiable citations from the web. This should lead to more accurate mental models of AI and its output — a research assistant, not an all-knowing oracle. This would already be an improvement over Google’s Knowledge Graph which frequently surfaces unsourced answers.
This is a demo inspired by OpenAI WebGPT, not a commercial product. Perplexity Ask is powered by large language models (OpenAI API) and search engines. Accuracy is limited by search results and AI capabilities.
YouChat is a ChatGPT-like AI search assistant that you can talk to right in your search results. It stays up-to-date with the news and cites its sources so that you can feel confident in its answers.
Vik Paruchuri‘s open source Researcher project:
By feeding web context into a large language model, you can improve accuracy and verify the information… Researcher gives you cited sources and more specific information by relying on context from Google.
It is getting increasingly perplexing that we haven’t yet seen any similar search demos from Google or Microsoft.
-
Ben Thompson, from his recent interview with Daniel Gross Nat Friedman:
And text is inherently a good match with deterministic thinking, because you can lay down explicitly what you mean. Yes, you don’t have the person’s internal monologue and thoughts, but they can articulate a whole lot of what is important and get that down on the page and walk you through their logic chain. And text lets you do that to a much greater sense than images. Images, a lot of it is your interpretation of the image. It’s you perceiving what it is. And interestingly, from a biological perspective, vision itself is probabilistic, right? That’s why you get those optical illusions, because your brain is filling in all the different pieces that go into it.
And this makes me wonder, maybe the real difference between deterministic computing and probabilistic computing is in fact the difference between text-based computing or text-based thinking and visual-based thinking. And this visual stuff is in fact not just a toy, it’s not just out there first because it’s easier to do. It actually might be the future of all of this.
-
Qwerty is going to be so sad when I take the tree down.
-
Remember how last week I mentioned that I successfully got through my final week of classes without getting sick? Well, almost immediately after hitting publish I started coming down with a cold. At least it wasn’t especially bad and I was back to feeling nearly 100% by Wednesday.
Anyway, this was my first week off this year and, despite the cold, I ended up accomplishing some of the household tasks I have been putting off for far too long. Most notably, painting a wall in my kitchen and deep cleaning the basement floor where it had been stained by cyanotype chemistry.
Links
- When Technology Changes Art
- Stable Diffusion architecture renders from dollhouse photos
- Related: Depth-to-image in Stable Diffusion
- Analog Diffusion - Stable Diffusion fine tuned on film photographs
- VectorFusion - text to SVG
- picoSYNTH - an interactive synthesizer running in the Pico-8 fantasy console
Recipes
- Tajín grilled chicken
- This was good but would have been better had I used an actual grill instead of the stovetop and broiler. Note to self to try it again next summer.
- Chile crisp fettuccine alfredo
- A better combination than I was expecting. I added lemon juice which was a good choice. Next time I make this I will probably try adding ground pork too.
- Thai carrot and sweet potato soup
- Eh, this was alright but definitely missing something although I am not sure I can pinpoint what exactly
-
Scott Aaronson, in a lecture at the University of Texas at Austin, describes a project he has been working on at OpenAI to watermark GPT output:
My main project so far has been a tool for statistically watermarking the outputs of a text model like GPT. Basically, whenever GPT generates some long text, we want there to be an otherwise unnoticeable secret signal in its choices of words, which you can use to prove later that, yes, this came from GPT.
At its core, GPT is constantly generating a probability distribution over the next token to generate… the OpenAI server then actually samples a token according to that distribution—or some modified version of the distribution, depending on a parameter called “temperature.” As long as the temperature is nonzero, though, there will usually be some randomness in the choice of the next token
So then to watermark, instead of selecting the next token randomly, the idea will be to select it pseudorandomly, using a cryptographic pseudorandom function, whose key is known only to OpenAI.
In my first two posts on the future of AI in education I highlighted proposals that frame generative AI use as an inevitability and, therefore, a tool to be embraced instead of a threat that warrants the development of technical detection mechanisms. While the risks of plagiarism and misinformation are undoubtedly real, we should push for a greater focus on strengthening critical analysis and fact-checking skills instead of starting an impossible to win arms race between detection and evasion.
The most exciting path forward is one where we frame Large Language Models as “a calculator for text”. Just as the invention of pocket calculators was a giant disruption that forced us to re-evaluate our approach to mathematics education, language models will continue to force us to re-evaluate our approach to research and writing. Done correctly this will open the door for us to learn more quickly, use our time more effectively, and progress further than we possibly could before.
-
I spent a good chunk of my day yesterday experimenting with Analog Diffusion — a Stable Diffusion model fine-tuned on film photography images. I have to say, I have been really impressed with almost everything this model has produced so far.
I can’t find any information online about which particular categories of images were part of the fine-tuning process so I started testing how the outcome would change when specifying different popular 35mm film stocks in the prompt:
The model certainly knows that Tri-X and HP5+ are both black and white films while Portra and Ektar are color films. Less clear to me is if there are any traits the model attributes to any single film stock in particular (for example, if images generated from “Ektar” prompts had more saturated colors than equivalent “Portra” prompt images).
Finally, a few open questions I would like to investigate soon:
- How well can the model emulate alternative photography processes like Daguerreotype, Collodion, and Tin Type?
- What about alternative development processes? Kodak D-76 vs. caffenol vs. stand development?
- Does specifying a high or low ISO have any perceivable effect?
- Do particular lenses and focal lengths behave as expected? How different are Summilux 35mm images from Summarit 120mm images?
-
Here is another thread related to what I wrote about yesterday. Even if text generation AI is primarily used to create boilerplate text, it can still be socially disruptive. There are a bunch of different categories of writing — letters of recommendation, job interview follow-ups, etc — where text, even if (or maybe because) it is largely formulaic, is used as an important social signal.
Kevin Munger, writing for Real Life magazine:
I can send a multi-line email thanking a friend for their birthday message, and neither of us can be sure to what degree the other was actually involved in the process. Such messages — despite potentially being identical to emails sent a decade ago — have now become more difficult to parse, not in terms of their information but their intention. Nothing can be reliably inferred from the fact my birthday was remembered or that I remembered to say thanks; no conclusions can be drawn from how timely the messages are.
Take letters of recommendation: the best letters are personal and heartfelt, but it has also become essential that they be long — vague claims about how great the student is are easy to come by, but a lengthy letter that contains specific details is a costly signal that the professor spent time and energy creating the letter. With GPT-3, however, it may become trivial for professors to plug in a CV and some details and end up with a lengthy, personalized letter. In the long run, this would undermine the value of such letters to credibly communicate effort and sincerity on the part of the recommender
-
"The fall (bababadalgharaghtakamminarronnkonnbronntonnerronntuonnthunntrovarrhounawnskawntoohoohoord-enenthurnuk!) of a once wallstrait oldparr is retaled early in bed and later on life down through all christian minstrelsy."— James Joyce, Finnegans Wake
Tyler Cowen writes that originality will become increasingly important as generative AI improves:
ChatGPT excels at producing ordinary, bureaucratic prose, written in an acceptable but non-descript style… rewards and status will go down for those who produce such writing today, and the rewards for exceptional originality are likely to rise. What exactly can you do to stand out from the froth of the chat bots?
There are two big axis of originality that you can optimize for: original ideas and original presentations. Being skilled at describing a familiar concept in a novel way, creating new analogies to more effectively explain old ideas, and the ability to generate entirely new ideas altogether will all continue to be highly valued skills.
Cowen goes on to say that some activities might be able to skip the competition for originality:
Alternatively, many humans will run away from such competitive struggles altogether. Currently the bots are much better at writing than say becoming a master gardener, which also requires skills of physical execution and moving in open space. We might thus see a great blossoming of talent in the area of gardening, and other hard to copy inputs, if only to protect one’s reputation and IP from the bots.
Athletes, in the broad sense of that term, may thus rise in status. Sculpture and dance might gain on writing in cultural import and creativity. Counterintuitively, if you wanted our culture to become more real and visceral in terms of what commands audience attention and inspiration, perhaps the bots are exactly what you’ve been looking for.
I still think that, when the dust settles, generative AI will be used primarily for filler: clip art, boilerplate, and advertising copy. Regardless, the most durable artworks will be those that are inbued with extratextual meaning. You’ll be in a better position if your work is closer to Kehinde Wiley than Mondrian.
-
Artist Agnieszka Pilat writing for Artnet:
Whenever a new medium is created, its context is an old medium. Shortly after the “electric wire panic,” the first feature films made at the beginning of the 20th-century used cameras and storylines to expand on the play; camera angles, non-chronological storytelling, and wide lenses came much later. With the advent of television, presenters initially stood in front of cameras and read the news as if it were a radio format.
Rather than assuming an outcome is preordained, artists should reclaim their agency and use A.I. in their work, becoming stakeholders in how the technology develops and directing its use in new mediums.
The introduction of new creative mediums typically follows the same pattern:
- Inventors, scientists, and engineers make a discovery
- Some subset of artists start experimenting with the new discovery
- Some subset of the public starts to panic and debate the definition of art
- Artists continue exploring the possible. First, by mimicking previous mediums and eventually carving out a new niche for the medium.
- The new medium gradually gains widespread acceptance in its new niche
New creative mediums don’t take off until they find their own niche. This can be a long, exciting process — one that we are living through now with AI generated art.
- The invention of photography didn’t simply make painting obsolete but instead forced artists to find the unique strengths of both mediums
- Just as movies aren’t simply recorded plays, an MP3 can be more than just a recorded live musical performance. You can layer tracks, speed them up, slow them down, chop and screw them, etc.
- A digital newspaper shouldn’t just be a static homepage when it can instead be a live, multimedia newsfeed
What will a true AI art medium look like? What will it let us create that we haven’t been able to create before?
-
Since first using ChatGPT I have been interested in how much improvement would be possible if:
- The network was able to extract pure mathematics from relevant questions and send those calculations off to Wolfram Alpha or a similar service
- The network was hooked into the internet for continuous live training and search to help answer factual questions
It turns out that at least that later question has been an ongoing topic of research for quite some time. From a December 2021 article published by OpenAI titled Improving the Factual Accuracy of Language Models through Web Browsing:
Language models like GPT-3 are useful for many different tasks, but have a tendency to “hallucinate” information when performing tasks requiring obscure real-world knowledge. To address this, we taught GPT-3 to use a text-based web-browser. The model is provided with an open-ended question and a summary of the browser state, and must issue commands such as “Search …”, “Find in page: …” or “Quote: …”. In this way, the model collects passages from web pages, and then uses these to compose an answer.
Our prototype copies how humans research answers to questions online—it submits search queries, follows links, and scrolls up and down web pages. It is trained to cite its sources, which makes it easier to give feedback to improve factual accuracy.
Even these early experiments appear to work surprisingly well with a roughly 3x increase in “truthful and informative” answers.
For more information on the specific prompts used in the graph above (“QA” vs “Helpful”) see Appendix E of TruthfulQA: Measuring How Models Mimic Human Falsehoods.
OpenAI notes that this approach does, of course, introduce new risks. Particularly generating answers using information from inaccurate sources.
Our models outperform GPT-3 on TruthfulQA and exhibit more favourable scaling properties. However, our models lag behind human performance, partly because they sometimes quote from unreliable sources
Eventually, having models cite their sources will not be enough to evaluate factual accuracy. A sufficiently capable model would cherry-pick sources it expects humans to find convincing, even if they do not reflect a fair assessment of the evidence.
However, if these risks are successfully mitigated, the resulting AI could not only be more factually accurate but also significantly smaller and less computationally intensive to run.
-
This was a strange week, as expected, with lots of student absences due to the flu, RSV, covid, and early vacations. Shockingly, I was able to avoid catching any of the illnesses myself and can now spend these next two weeks catching up on things at home and, I hope, relaxing. This is really my first chance to expirence the perks of a teacher’s schedule and I don’t intend to take it for granted.
It snowed for the first time this year overnight Friday. The holiday season has truly begun.
Links
- Disputing a parking fine with ChatGPT
- Notice, also, how the response to the author’s email was almost certainly auto-generated / boilerplate
- AI sculpting
- More technical details here
- A list of indie Mac and iOS applications that promote the open web
- Qualia Research Institute’s first line of scents
- How does GPT Obtain its Ability?
- Using Stable Diffusion to generate music
- Textcasting
- Disputing a parking fine with ChatGPT
-
Zeynep Tufekci, writing for the New York Times, presents another way generative AI tools can be embraced in a classroom setting:
Schools have already been dealing with the internet’s wealth of knowledge, along with its lies, misleading claims and essay mills.
One way has been to change how they teach. Rather than listen to a lecture in class and then go home to research and write an essay, students listen to recorded lectures and do research at home, then write essays in class, with supervision, even collaboration with peers and teachers. This approach is called flipping the classroom.
In flipped classrooms, students wouldn’t use ChatGPT to conjure up a whole essay. Instead, they’d use it as a tool to generate critically examined building blocks of essays. It would be similar to how students in advanced math classes are allowed to use calculators to solve complex equations without replicating tedious, previously mastered steps.
Teachers could assign a complicated topic and allow students to use such tools as part of their research. Assessing the veracity and reliability of these A.I.-generated notes and using them to create an essay would be done in the classroom, with guidance and instruction from teachers. The goal would be to increase the quality and the complexity of the argument.
This idea is similar to Ben Thompson’s proposal that I wrote about last week. While still not perfect, Tufekci’s plan at least wouldn’t require the use of bespoke educational technologies that would necessarily lag behind state of the art general-purpose alternatives.
-
Eryk Salvaggio compares the process of creating with generative AI to creative play inside of a video game:
Video games create systems that shape the way we play them. AI art interfaces create structures that allow us to play. These interfaces offer genuine delight and curiosity. But like video games, this delight and curiosity is bounded by invisible mechanics, which ultimately shape the desires that we bring into them.
It is possible to create genuinely interesting things within the constraints of a video game. You are, however, still limited by what the game designers originally intended — you can’t do anything fundamentally new because everything you do must first be programmed into the game itself.
Because of these limitations, Salaggio argues that artists need to completely reimagine their relationship with AI tools just as Cory Arcangel did when he hacked a Super Mario Brothers cartridge to remove everything other than the clouds.
How we do move our relationship, as artists using DALLE2, away from consuming its choices, and into producing work — to steer this relationship away from machine dominance and toward our own desires? In other words: how do we begin to make our own games, rather than playing somebody else’s?
The most exciting possibilities for creating AI art are not in the nuance or fascinating minutiae of prompt engineering. The potential of AI art is yet to be realized. It will come about as artists learn to pull away from the call-and-response of the prompt window and its prizes.
When we defend tools like DALLE2 as if the raw product is an artistic expression of the one who types the prompt, we sacrifice some of that criticality in negotiating the lines between personal expression and technologically mediated desire. The prompt is an idea, but the image — the expression of that idea — is left to the machine.
Let’s embrace these outputs as a source of further play, not the end point.
Midjourney, which was, until very recently, built on top of the open source Stable Diffusion model, has an intentionally distinct style that is consistent across all of the images it produces. This is a deliberate choice the Midjourney developers made because it is a differentiator: the reason you would pay for Midjourney instead of using a free alternative is because you want the unique, proprietary style that Midjourney offers.
This might serve another purpose too. An AI image generator that has a recognizable “house style” is obviously not a neutral medium. It is not simply an unbiased extension of the artist but a tool with constraints, not fully under your control.
-
Stanislas Polu has some predictions about the state of large language models next year. The only prediction of his that I feel especially confident in myself is the following:
There will be an opensource Chinchilla-style LLM released this year at the level of text-davinci-*. Maybe not from the ones we expect🤔This will obliterate ChatGPT usage and enable various types of fine-tuning / soft-prompting and cost/speed improvements.
I would say that now, especially after the success of ChatGPT, an equivalent open source LLM will almost certainly be released in the next year. This will likely follow the same pattern as image generation AIs earlier this year: first, OpenAI released DALL-E 2 as a private beta. Then, a little while later, Stable Diffusion was released which, although it wasn’t quite as good as DALL-E, was free, open-source, and widely accessible. This allowed for an explosion of creative applications including photoshop plugins, 3D modeling plugins, and easy to install native frontend interfaces.
While I believe text generation AIs will have a similar moment early next year. The unfortunate truth is that running even a pre-trained text generation network requires significantly more computer memory than is required to run similarly sized image generation networks. This means we will probably not see early open-source text generation networks running natively on consumer hardware such as iPhones like we have with Stable Diffusion (although it is possible Apple will, once again, help with that).
My hope is that these developments in text-generation spur some much-needed innovation in household voice assistants which are increasingly feeling dated.
-
Onformative Studio shares its expirence using a generative AI to create 3D sculptures:
We were guided by the question of how our role as the designer is changing when human creativity and artificial intelligence co-create. In the course of the AI’s sculpting process we were inspired by the unpredictable strategies and outcomes of the reinforcement learning: an experimental approach par excellence, which we guided, observed and visualized.
Above all, we have also questioned our own role as creators. Rather than leaving creation to AI, we need to find ways to integrate it into the creative process. We take technology as a starting point, a tool, a source of inspiration and a creative partner. The human aspect is quite clear in this: We choose the rules and define the approximate output. However, in the end it was the interplay between our human choices and the agent’s ability to find the best solutions and occasionally surprise us. This made the process rewarding to us and shows the true potential of an AI based co-creation process.
This reminds me of some of the ideas Noah Smith and Roon laid out in their recent article about generative AI:
We think that the work that generative AI does will basically be “autocomplete for everything”.
What’s common to all of these visions is something we call the “sandwich” workflow. This is a three-step process. First, a human has a creative impulse, and gives the AI a prompt. The AI then generates a menu of options. The human then chooses an option, edits it, and adds any touches they like.
There’s a natural worry that prompting and editing are inherently less creative and fun than generating ideas yourself, and that this will make jobs more rote and mechanical.
Ultimately, though, we predict that lots of people will just change the way they think about individual creativity. Just as some modern sculptors use machine tools, and some modern artists use 3d rendering software, we think that some of the creators of the future will learn to see generative AI as just another tool – something that enhances creativity by freeing up human beings to think about different aspects of the creation.
I am optimistic that generative AI will continue to make creative expression of all kinds more accessible. At first, AI assisted design will be viewed as lower status or less authentic, like baking brownies using a pre-made box mix instead of “from scratch.” Later, though, once these technologies become more established, not using an AI will prompt discussion and warrant explanation — AI collaboration will be the norm. This is already the case with smartphone photography.
-
Matt Yglesias on how technological progress might change higher education:
The past couple of centuries have seen a steady increase in the market demand for certain kinds of skills. That’s meant people want to acquire these skills even if they don’t necessarily have tons of inherent motivation — they want training to earn a better living, and education as a formal process is a very useful motivational crutch. It’s at least possible that the next twist in IT will be to rapidly erode the demand for that kind of education, meaning people will be left primarily to learn things they are curious about where there is much less need for external motivation.
This is part of the reason why I believe fostering a strong and diverse set of intrinsic motivations is crucial at all levels of education. Whether or not the future Yglesias describes here comes to pass, students driven by intrinsic motivations are always going to be more passionate and engaged than students driven by extrinsic motivations.
-
Sam Kriss on how, as GPT models have become more accurate, they have become less compelling:
I tried using GPT-2 to write a novel. I let the AI choose the title, and it was absolutely insistent that it should be called “BONKERS FROM MY SLEEVE”, with the caps and quotation marks very much included. There was a Pynchonian array of bizzarely named characters (including the Birthday Skeletal Oddity, Thomas the Fishfaller, the Hideous Mien of Lesbian Jean, a ‘houndspicious’ Labradoodle named Bam Bam, and a neo-Nazi cult called ‘The Royal House of the Sun’), none of the episodes made any sense whatsoever, and major characters had a habit of abruptly dying and then popping up again with a different gender. But along the way it produced some incredible passages
…
GPT-2 was not intended as a machine for being silly, but that’s what it was—precisely because it wasn’t actually very good at generating ordinary text. The vast potential of all possible forms kept on seeping in around the edges, which is how it could generate such strange and beautiful strings of text. But GPT-3 and its applications have managed to close the lid on chaos.
…
There are plenty of funny ChatGPT screenshots floating around. But they’re funny because a human being has given the machine a funny prompt, and not because the machine has actually done anything particularly inventive. The fun and the comedy comes from the totally straight face with which the machine gives you the history of the Toothpaste Trojan War. But if you gave ChatGPT the freedom to plan out a novel, it would be a boring, formulaic novel, with tedious characters called Tim and Bob, a tight conventional plot, and a nice moral lesson at the end. GPT-4 is set to be released next year. A prediction: The more technologically advanced an AI becomes, the less likely it is to produce anything of artistic worth.
I could also see this as something that would help establish the validity of AI assisted artwork. As in “AI could never make something this inventive on its own, the artist must have played a large role in the artwork’s creation.”
-
Next week is the last week before winter break, not that I am counting.
I made some homemade pizza in the outdoor pizza oven on Saturday. For the first time my gluten-free dough tasted the same as my glutenous dough. I hope this means my gluten-free skills are improving and not that I simply don’t know how to cook with gluten anymore. Anyway, the pizza was good and I don’t think I will have many more opprotunites to use the pizza oven until spring.
Links
- Awesome ChatGPT Prompts
- This list also includes prompts written by ChatGPT itself
- Ooh.directory — A collection of 1000+ personal blogs. Run by Phil Gyford who, unsuprisingly, has his own personal blog that I enjoy.
- See also: The Mataroa Collection
- Obsolete Sounds
- Introduction to Genomics for Engineers
- Awesome ChatGPT Prompts
-
I teach a technology class to students from Kindergarten to fifth grade. The accelerated development of truly impressive AI models recently — especially ChatGPT and Stable Diffusion — has made it clear to me how dramatically different technological literacy will be when my students eventually enter the world as adults.
As we move into a future with increased technological automation, forward-looking curricula across all subject matters must focus on fostering creativity in students. Although AI can make new connections between elements in its training data, it is humans alone that are capable of generating truly novel ideas.
I believe teaching young students how to code will continue to be important to some extent. However, with the rise of code generation technologies like GitHub CoPilot, the most durable programming skills might be spec-writing, debugging, and revising. Physical electronics and robotics will arguably rise in relevance for the foreseeable future. Understanding and debugging systems will be an important skill here, too.
It would be great to hear from other educators that are thinking through similar topics right now.
-
Zero Trust Homework
Here’s an example of what homework might look like under this new paradigm. Imagine that a school acquires an AI software suite that students are expected to use for their answers about Hobbes or anything else; every answer that is generated is recorded so that teachers can instantly ascertain that students didn’t use a different system. Moreover, instead of futilely demanding that students write essays themselves, teachers insist on AI. Here’s the thing, though: the system will frequently give the wrong answers (and not just on accident — wrong answers will be often pushed out on purpose); the real skill in the homework assignment will be in verifying the answers the system churns out — learning how to be a verifier and an editor, instead of a regurgitator.
I am not sure I fully agree with Ben’s proposal here but, at the same time, I am having trouble coming up with any coherent solutions for homework / assessments that truly account for the AI we have today — let alone what we will have 10 years from now.
Ultimately, I am hopeful these advances in AI will push us to re-evaluate our current approach to education and lead us towards more meaningful, interdisciplinary, and practical approaches in the future.
subscribe via RSS