Daily Archives: November 1, 2024
OpenAI’s search engine is now live in ChatGPT
ChatGPT is officially an AI-powered web search engine. The company is enabling real-time information in conversations for paid subscribers today (along with SearchGPT waitlist users), with free, enterprise, and education users gaining access in the coming weeks.
Rather than launching as a separate product, web search will be integrated into ChatGPT’s existing interface. The feature determines when to tap into web results based on queries, though users can also manually trigger web searches. ChatGPT’s web search integration finally closes a key competitive gap with rivals like Microsoft Copilot and Google Gemini, which have long offered real-time internet access in their AI conversations. — Read More
Scaling and evaluating sparse autoencoders
Sparse autoencoders provide a promising unsupervised approach for extracting interpretable features from a language model by reconstructing activations from a sparse bottleneck layer. Since language models learn many concepts, autoencoders need to be very large to recover all relevant features. However, studying the properties of autoencoder scaling is difficult due to the need to balance reconstruction and sparsity objectives and the presence of dead latents. We propose using k-sparse autoencoders [Makhzani and Frey, 2013] to directly control sparsity, simplifying tuning and improving the reconstruction-sparsity frontier. Additionally, we find modifications that result in few dead latents, even at the largest scales we tried. Using these techniques, we find clean scaling laws with respect to autoencoder size and sparsity. We also introduce several new metrics for evaluating feature quality based on the recovery of hypothesized features, the explainability of activation patterns, and the sparsity of downstream effects. These metrics all generally improve with autoencoder size. To demonstrate the scalability of our approach, we train a 16 million latent autoencoder on GPT-4 activations for 40 billion tokens. We release training code and autoencoders for open-source models, as well as a visualizer. — Read More
Evaluating feature steering: A case study in mitigating social biases
A few months ago, we published an interpretability paper demonstrating our ability to learn interpretable features that correspond to various concepts (e.g., famous individuals, types of computer code, etc.) represented in Claude 3 Sonnet. To verify our feature interpretations, we ran qualitative feature steering experiments, where we artificially dialed up and down various features to see if they changed model outputs in intuitive ways. The results were promising – for example, turning up a feature that responded to mentions of the Golden Gate Bridge made the model talk about the Golden Gate Bridge. Such examples led us to hypothesize that feature steering might be a promising way to modify model outputs in specific interpretable ways. — Read More
You could start smelling the roses from far away using AI
AI can “teleport” scents without human hands (or noses)
Ever send a picture of yourself trying on clothes to a friend to see what they think of how you look? Now, imagine doing the same from the perfume and cologne counter. AI could make that happen in the not-too-distant future after a breakthrough in ‘Scent Teleportation.’ Osmo, which bills itself as a “digital olfaction” company, has succeeded in using AI to analyze a scent in one location and reproduce it elsewhere without human intervention. — Read More