How to make music with AI using Udio

There’s something quite alluring about trying to create art in a form you’re less familiar with. AI music is the latest canvas in this space.

While we can easily sketch a drawing with a pen and piece of paper at home, not all of us have instruments lying around or the skills to use them.

Generative AI gets rid of those hurdles and tools like Udio, Stable Audio, Cassette AI and Suno allow us to dip our toes into music production. Prior experience is not required. Furthermore, Udio seems to be on to something in that it is able to combine a simple user experience with pretty decent results. — Read More

#audio

Introducing more enterprise-grade features for API customers

We[OpenAI]’ve introduced Private Link, a new way that customers can ensure direct communication between Azure and OpenAI while minimizing exposure to the open internet. We’ve also released native Multi-Factor Authentication (MFA) to help ensure compliance with increasing access control requirements. These are new additions to our existing stack of enterprise security features including SOC 2 Type II certification, single sign-on (SSO), data encryption at rest using AES-256 and in transit using TLS 1.2, and role-based access controls. We also offer Business Associate Agreements for healthcare companies that require HIPAA compliance and a zero data retention policy for API customers with a qualifying use case. — Read More

#cyber

Microsoft launches Phi-3, its smallest AI model yet

Microsoft launched the next version of its lightweight AI model Phi-3 Mini, the first of three small models the company plans to release. 

Phi-3 Mini measures 3.8 billion parameters and is trained on a data set that is smaller relative to large language models like GPT-4. It is now available on Azure, Hugging Face, and Ollama. Microsoft plans to release Phi-3 Small (7B parameters) and Phi-3 Medium (14B parameters). Parameters refer to how many complex instructions a model can understand.  — Read More

#nlp

AI Is Turning into Something Totally New — Mustafa Suleyman — TED

Read More

#videos

Drake Uses AI Tupac and Snoop Dogg Vocals on ‘Taylor Made Freestyle

The beef between Drake and what continues to be a strong sect of the hip-hop community grows deeper. On Friday night (April 19), the rapper released a song on his social media entitled “Taylor Made Freestyle,” which uses AI vocals from Tupac Shakur and Snoop Dogg on a stopgap between diss records as he awaits Kendrick Lamar’s reply to his freshly released “Push Ups.”Read More

#audio

How Meta is paving the way for synthetic social networks

On Thursday, the AI hype train rolled through Meta’s family of apps. The company’s Meta AI assistant, a ChatGPT-like bot that can answer a wide range of questions, is beginning to roll out broadly across Facebook, Messenger, Instagram and WhatsApp.

Powering the bot is Llama 3, the latest and most capable version of Meta’s large language model. As with its predecessors — and in contrast to models from OpenAI, Google, and Anthropic — Llama 3 is open source. Today Meta made it available in two sizes: one with 8 billion parameters, and one with 70 billion parameters. (Parameters are the variables inside a large language model; in general, the more parameters a model contains, the smarter and more sophisticated its output.) — Read More

#big7, #devops

Maybe I don’t want a Rosey the Robot after all

Boston Dynamics’ latest — deliberately creepy? — humanoid robot has me rethinking my smart home robot dreams.

As a child of the 1980s, my perception of the smart home has been dominated by the idea that one day, we will all have Rosey the Robot-style robots roaming our homes — dusting the mantelpiece, preparing dinner, and unloading the dishwasher. (That last one is a must; we were smart enough to come up with a robot to wash our dishes; can’t we please come up with one that can also unload them?)

However, after seeing Boston Dynamics’ latest droid, Atlas, unveiled this week, my childhood dreams are fast turning into a smart home nightmare. While The Jetsons’ robot housekeeper had a steely charm, accentuated by its frilly apron, the closer we come to having humanoid robots in our home, the more terrifying it appears they will be. Not so much because of how they look — I could see Atlas in an apron — but more because of what they represent.  — Read More

#robotics

The Universal Approximation Theorem

Artificial Intelligence has become very present in the media in the last couple of years. At the end of 2022, ChatGPT has captured the world’s attention, showing at least a hundred million users around the globe the extraordinary potential of large language models. Large language models such as LLaMABard and ChatGPT mimic intelligent behavior equivalent to, or indistinguishable from, that of a human in specific areas (i.e., Imitation Game or Turing Test). Stephen Wolfram has written an article about how ChatGPT works.

Year-end 2022 might therefore be a watershed moment for human mankind since Artificial Intelligence has now the potential to change the way how humans think and work

… All these achievements have one thing in common – they are build on a model using an Artificial Neural Networks (ANN). … ANN are very good function approximators provided that big data of the corresponding domain is available. Almost all practical problems such as playing a game of Go or mimic intelligent behavior can be represented by mathematical functions.

The corresponding theorem that formulates this basic idea of approximation is called Universal Approximation Theorem. It is a fundamental result in the field of ANN, which states that certain types of neural network can approximate certain function to any desired degree of accuracy. This theorem suggest that a neural network is capable of learning complex patterns and relationships in data as long as certain conditions are fulfilled.

The Universal Approximation Theorem is the root-cause why ANN are so successful and capable in solving a wide range of problems in machine learning and other fields. — Read More

#deep-learning

Microsoft AI creates scary real talkie videos from a single photo

Microsoft Research Asia has revealed an AI model that can generate frighteningly realistic deepfake videos from a single still image and an audio track. How will we be able to trust what we see and hear online from here on in?

… After training the [VASA-1] model on footage of around 6,000 real-life talking faces from the VoxCeleb2 dataset, the technology is able to generate scary real video where the newly animated subject is not only able to accurately lip-sync to a supplied voice audio track, but also sports varied facial expressions and natural head movements – all from a single static headshot photo. — Read More

#image-recognition

Google Consolidates AI-Building Teams Across Research and DeepMind

Google is consolidating the teams that focus on building artificial intelligence (AI) models across Google Research and Google DeepMind

All this work will now be done within Google DeepMind, Sundar Pichai, CEO of Google and Alphabet, said in a note to employees posted on the company’s website Thursday (April 18). — Read More

#big7, #strategy