gpt-author

This project utilizes a chain of GPT-4 and Stable Diffusion API calls to generate an original fantasy novel. Users can provide an initial prompt and enter how many chapters they’d like it to be, and the AI then generates an entire novel, outputting an EPUB file compatible with e-book readers.

A 15-chapter novel can cost as little as $4 to produce, and is written in just a few minutes. — Read More

#devops, #multi-modal

Project S.A.T.U.R.D.A.Y — A Vocal Computing Toolbox

A toolbox for vocal computing built with Pionwhisper.cpp, and Coqui TTS. Build your own personal, self-hosted J.A.R.V.I.S powered by WebRTC

Project S.A.T.U.R.D.A.Y is a toolbox for vocal computing. It provides tools to build elegant vocal interfaces to modern LLMs. The goal of this project is to foster a community of like minded individuals who want to bring forth the technology we have been promised in sci-fi movies for decades. It aims to be highly modular and flexible while staying decoupled from specific AI Models. This allows for seamless upgrades when new AI technology is released. — Read More

#audio, #devops

What is Langchain and why should I care as a developer?

Langchain is one of the fastest growing open source projects in history, in large part due to the explosion of interest in LLM’s.

This post explores some of the cool thing that langchain helps developers do from a 30,000 foot overview. It was written for my own benefit as I explored the framework and I hope it helps you if you are also curios where langchain might be useful.

Some of the features that make langchain so powerful include allowing you to connect data to language models (like OpenAI’s GPT models via the API) and create agent workflows (more on agents later). — Read More

#devops

Run open-source LLMs on your computer. Works offline. Zero configuration.

Discover the remarkable capabilities of open-source LLMs on your personal computer. Operate seamlessly without an internet connection and with effortless setup. — Read More

#chatbots, #devops

The New Language Model Stack

ChatGPT unleashed a tidal wave of innovation with large language models (LLMs). More companies than ever before are bringing the power of natural language interaction to their products. The adoption of language model APIs is creating a new stack in its wake. To better understand the applications people are building and the stacks they are using to do so, we spoke with 33 companies across the Sequoia network, from seed stage startups to large public enterprises. We spoke with them two months ago and last week to capture the pace of change. As many founders and builders are in the midst of figuring out their AI strategies themselves, we wanted to share our findings even as this space is rapidly evolving.  — Read More

#devops

Leveraging FastAPI, OpenAI, and SQLAlchemy for Natural Language SQL Queries

SQL (Structured Query Language) is the standard language for managing and manipulating relational databases. What if we could interact with databases using natural language queries?

In this post we show how you can use SQL to load a dataframe to a database, write a prompt to query it, and connect this to a FastAPI application for deployment and enabling users to interact with the database. — Read More

#devops

How open-source LLMs are challenging OpenAI, Google, and Microsoft

In the past few years, it seemed that wealthy tech companies would be able to monopolize the growing market for large language models (LLM). And recent earnings calls from big tech companies suggested they are in control. Microsoft’s announcements, in particular, show that the company has created a billion-dollar business from its AI services, including through Azure OpenAI Services and the workloads OpenAI runs on its cloud infrastructure.

However, a recently leaked internal document from Google indicates that the market share of big tech is not as secure as it seems thanks to advances in open-source LLMs. In short, the document says “We have no moat, and neither does OpenAI.” The dynamics of the market are gradually shifting from “bigger is better” to “cheaper is better,” “more efficient is better,” and “customizable is better.” And while there will always be a market for cloud-based LLM and generative AI products, customers now have open-source options to explore as well. — Read More

#devops, #nlp

Google DeepMind’s game-playing AI just found another way to make code faster

The AI-generated algorithms are already being used by millions of developers.

DeepMind’s run of discoveries in fundamental computer science continues. Last year the company used a version of its game-playing AI AlphaZero to find new ways to speed up the calculation of a crucial piece of math at the heart of many different kinds of code, beating a 50-year-old record.

Now it has pulled the same trick again—twice. Using a new version of AlphaZero called AlphaDev, the UK-based firm (recently renamed Google DeepMind after a merge with its sister company’s AI lab in April) has discovered a way to sort items in a list up to 70% faster than the best existing method.

It has also found a way to speed up a key algorithm used in cryptography by 30%. These algorithms are among the most common building blocks in software. Small speed-ups can make a huge difference, cutting costs and saving energy. — Read More

Read the Paper

#devops

RedPajama 7B now available, instruct model outperforms all open 7B models on HELM benchmarks

The RedPajama project aims to create a set of leading open-source models and to rigorously understand the ingredients that yield good performance. In April we released the RedPajama base dataset based on the LLaMA paper, which has worked to kindle rapid innovation in open-source AI.

The 5 terabyte dataset has been downloaded thousands of times and used to train over 100 models! Read More

#chatbots, #devops

The Falcon has landed in the Hugging Face ecosystem

Falcon is a new family of state-of-the-art language models created by the Technology Innovation Institute in Abu Dhabi, and released under the Apache 2.0 license. Notably, Falcon-40B is the first “truly open” model with capabilities rivaling many current closed-source models. This is fantastic news for practitioners, enthusiasts, and industry, as it opens the door for many exciting use cases.

In this blog, we will be taking a deep dive into the Falcon models: first discussing what makes them unique and then showcasing how easy it is to build on top of them (inference, quantization, finetuning, and more) with tools from the Hugging Face ecosystem. — Read More

#devops, #nlp