A good show is going on.
The Industrial Internet is a word that sounds quite unfamiliar to ordinary people and a little unsophisticated in technology circles.
Rooted in industry, it doesn’t sound as cool as metaverse sci-fi, and it doesn’t have the immediate effect of other high-tech applications.
However, the Industrial Internet is actually everywhere – as big as China’s international strategic position, as small as you and me. It stands there prominently among the seven main items of the “new infrastructure” concept, and it has become a celebrity in China’s macro and micro strategies for four consecutive years. Read More
Monthly Archives: February 2022
CSLIM Brings AI-Generated Art to ‘The World,’ or Vice Versa
Welcome to the world of tomorrow, where artificial intelligence can create East Asian landscape paintings that rival 11th century masters’ — with ownership that can be verified on the blockchain. Such is the vision of South Korean generative artist CSLIM, whose upcoming NFT collection will feature 5,000 artworks that the engineer and developer attests were created by AI that has studied the data of 80,000 classical East Asian landscape paintings from the sixth through 13th centuries via machine learning. Exquisitely combining cultural tradition and cutting-edge technology, “The World” will drop on Feb. 21 — only at Crypto.com/NFT. Read More
I asked an AI to paint 10 famous sci-fi book titles in one minute. Here are the results.
Fifty years ago, top scientists believed AI would never be able to beat humans at chess.
We all know how that turned out.
The newest goalpost is art. As an AI artist, people routinely tell me how my art isn’t real, or how it lacks humanity because it’s machine generated.
True, there’s some credence to these claims. But I would venture most people are simply uncomfortable with the notion that AI is starting to produce art faster & better than humans.
After my last post on AI art blew up, it naturally attracted a fair amount of similar criticism. Much was in reference to a supposed lack of diversity in the samples (apparently all of the landscapes looked the same).
To respond to this criticism, I decided to run an A100 GPU for one minute. Read More
Good News About the Carbon Footprint of Machine Learning Training
Machine learning (ML) has become prominent in information technology, which has led some to raise concerns about the associated rise in the costs of computation, primarily the carbon footprint, i.e., total greenhouse gas emissions. While these assertions rightfully elevated the discussion around carbon emissions in ML, they also highlight the need for accurate data to assess true carbon footprint, which can help identify strategies to mitigate carbon emission in ML.
In “The Carbon Footprint of Machine Learning Training Will Plateau, Then Shrink”, accepted for publication in IEEE Computer, we focus on operational carbon emissions — i.e., the energy cost of operating ML hardware, including data center overheads — from training of natural language processing (NLP) models and investigate best practices that could reduce the carbon footprint. We demonstrate four key practices that reduce the carbon (and energy) footprint of ML workloads by large margins, which we have employed to help keep ML under 15% of Google’s total energy use. Read More
Best of arXiv — February 2022
A monthly selection of ML papers by Zeta Alpha: Reinforcement Learning, Multimodality, Language Models as a service, Computer Vision, Information Retrieval and more. Read More
A replay of life: What happens in our brain when we die?
Neuroscientists have recorded the activity of a dying human brain and discovered rhythmic brain wave patterns around the time of death that are similar to those occurring during dreaming, memory recall, and meditation. Now, a study published to Frontiers brings new insight into a possible organizational role of the brain during death and suggests an explanation for vivid life recall in near-death experiences.
Imagine reliving your entire life in the space of seconds. Like a flash of lightning, you are outside of your body, watching memorable moments you lived through. This process, known as ‘life recall’, can be similar to what it’s like to have a near-death experience. What happens inside your brain during these experiences and after death are questions that have puzzled neuroscientists for centuries. However, a new study published to Frontiers in Aging Neuroscience suggests that your brain may remain active and coordinated during and even after the transition to death, and be programmed to orchestrate the whole ordeal.
When an 87-year-old patient developed epilepsy, Dr Raul Vicente of the University of Tartu, Estonia and colleagues used continuous electroencephalography (EEG) to detect the seizures and treat the patient. During these recordings, the patient had a heart attack and passed away. This unexpected event allowed the scientists to record the activity of a dying human brain for the first time ever. Read More
People Trust Deepfake Faces Generated by AI More Than Real Ones, Study Finds
The proliferation of deepfake technology is raising concerns that AI could start to warp our sense of shared reality. New research suggests AI-synthesized faces don’t simply dupe us into thinking they’re real people, we actually trust them more than our fellow humans.
In 2018, Nvidia wowed the world with an AI that could churn out ultra-realistic photos of people that don’t exist. Its researchers relied on a type of algorithm known as a generative adversarial network (GAN), which pits two neural networks against each other, one trying to spot fakes and the other trying to generate more convincing ones. Given enough time, GANS can generate remarkably good counterfeits.
Since then, capabilities have improved considerably, with some worrying implications: enabling scammers to trick people, making it possible to splice people into porn movies without their consent, and undermining trust in online media. While it’s possible to use AI itself to spot deepfakes, tech companies’ failures to effectively moderate much less complicated material suggests this won’t be a silver bullet. Read More
ET-BERT: A Contextualized Datagram Representation with Pre-training Transformers for Encrypted Traffic Classification
Encrypted traffic classification requires discriminative and robust traffic representation captured from content-invisible and imbalanced traffic data for accurate classification, which is challenging but indispensable to achieve network security and network management. The major limitation of existing solutions is that they highly rely on the deep features, which are overly dependent on data size and hard to generalize on unseen data. How to leverage the open-domain unlabeled traffic data to learn representation with strong generalization ability remains a key challenge. In this paper, we propose a new traffic representation model called Encrypted Traffic Bidirectional Encoder Representations from Transformer (ET-BERT), which pre-trains deep contextualized datagram-level representation from large-scale unlabeled data. The pre-trained model can be fine-tuned on a small number of task-specific labeled data and achieves state-of-the-art performance across five encrypted traffic classification tasks, remarkably pushing the F1 of ISCX-VPN-Service to 98.9% (5.2%↑), Cross-Platform (Android) to 92.5% (5.4%↑), CSTNET-TLS 1.3 to 97.4% (10.0%↑). Notably, we provide explanation of the empirically powerful pre-training model by analyzing the randomness of ciphers. It gives us insights in understanding the boundary of classification ability over encrypted traffic. The code is available at: https://github.com/linwhitehat/ET-BERT Read More
#nlpFewNLU: Benchmarking State-of-the-Art Methods for Few-Shot Natural Language Understanding
The few-shot natural language understanding (NLU) task has attracted much recent attention. However, prior methods have been evaluated under a disparate set of protocols, which hinders fair comparison and measuring progress of the field. To address this issue, we introduce an evaluation framework that improves previous evaluation procedures in three key aspects, i.e., test performance, dev test correlation, and stability. Under this new evaluation framework, we re-evaluate several state-of-the-art few-shot methods for NLU tasks. Our framework reveals new insights: (1) both the absolute performance and relative gap of the methods were not accurately estimated in prior literature; (2) no single method dominates most tasks with consistent performance; (3) improvements of some methods diminish with a larger pretrained model; and (4) gains from different methods are often complementary and the best combined model performs close to a strong fully-supervised baseline. We open-source our toolkit, FewNLU, that implements our evaluation framework along with a number of state-of-the-art methods. Read More
#nlp, #performanceCompute Trends Across Three Eras of Machine Learning
Compute, data, and algorithmic advances are the three fundamental factors that guide the progress of modern Machine Learning (ML). In this paper we study trends in the most readily quantified factor – compute. We show that before 2010 training compute grew in line with Moore’s law, doubling roughly every 20 months. Since the advent of Deep Learning in the early 2010s, the scaling of training compute has accelerated, doubling approximately every 6 months. In late 2015, a new trend emerged as firms developed large-scale ML models with 10 to 100-fold larger requirements in training compute. Based on these observations we split the history of compute in ML into three eras: the Pre Deep Learning Era , the Deep Learning Era and the Large-Scale Era . Overall, our work highlights the fast-growing compute requirements for training advanced ML systems. Read More