Imagine…imagine that you have been challenged to play Steph Curry, the greatest 3-point shooter in the history of the National Basketball Association, in a game of 1×1. Yea, a pretty predictable outcome for 99.9999999% of us.
But now image that Steph Curry has to wear a suit of knight’s armor as part of that 1×1 game. The added weight, the obstructed vision, and the lack of flexibility, agility and mobility would probably allow even the average basketball player to beat him.
Welcome to today’s technology architecture challenge! Read More
Daily Archives: June 20, 2019
Challenges and opportunities with Computer Vision at the Edge
IBM’s AI automatically generates creative captions for images
Writing photo captions is a monotonous — but necessary — chore begrudgingly undertaken by editors everywhere. Fortunately for them, though, AI might soon be able to handle the bulk of the work. In a paper (“Adversarial Semantic Alignment for Improved Image Captions”) appearing at the 2019 Conference in Computer Vision and Pattern Recognition (CVPR) in Long Beach, California this week, a team of scientists at IBM Research describes a model capable of autonomously crafting diverse, creative, and convincingly humanlike captions. Read More
Adversarial Semantic Alignment for Improved Image Captions
In this paper we study image captioning as a conditional GAN training, proposing both a context-aware LSTM captioner and co-attentive discriminator, which enforces semantic alignment between images and captions. We empirically focus on the viability of two training methods: Self-critical Sequence Training (SCST) and Gumbel Straight-Through (ST) and demonstrate that SCST shows more stable gradient behavior and improved results over Gumbel ST, even without accessing discriminator gradients directly. We also address the problem of automatic evaluation for captioning models and introduce a new semantic score, and show its correlation to human judgement. As an evaluation paradigm, we argue that an important criterion for a captioner is the ability to generalize to compositions of objects that do not usually cooccur together. To this end, we introduce a small captioned Out of Context (OOC) test set. The OOC set, combined with our semantic score, are the proposed new diagnosis tools for the captioning community. When evaluated on OOC and MS-COCO benchmarks, we show that SCST-based training has a strong performance in both semantic score and human evaluation, promising to be a valuable new approach for efficient discrete GAN training. Read More
Automated Machine Learning (AML) Comes of Age – Almost
You know you’ve come of age when the major review publications like Gartner and Forrester publish a study on your segment. That’s what’s finally happened. Just released is “The Forrester New Wave™: Automation-Focused Machine Learning Solutions, Q2 2019”.
This is the first reasonably deep review of platforms and covers nine of what Forrester describes as ‘the most significant providers in the segment’. Those being Aible, Bell Integrator, Big Squid, DataRobot, DMway Analytics, dotData, EdgeVerve, H2O.ai, and Squark.
I’ve been following these automated machine learning (AML) platforms since they emerged. I wrote first about them in the spring of 2016 under the somewhat scary title “Data Scientists Automated and Unemployed by 2025!”.
Well we’ve still got six years to run and it hasn’t happened yet. On the other hand no-code data science is on the rise and AML platforms along with their partially automated platform brethren are what’s behind it. Read More
World Models
We explore building generative neural network models of popular reinforcement learning environments. Our world model can be trained quickly in an unsupervised manner to learn a compressed spatial and temporal representation of the environment. By using features extracted from the world model as inputs to an agent, we can train a very compact and simple policy that can solve the required task. We can even train our agent entirely inside of its own hallucinated dream generated by its world model, and transfer this policy back into the actual environment. Read More
This AI Learns From Its Dreams
Linking Artificial Intelligence Principles
Various Artificial Intelligence Principles are designed with different considerations, and none of them can be perfect and complete for every scenario. Linking Artificial Intelligence Principles (LAIP)is an initiative and platform for synthesizing, linking, and analyzing various Artificial Intelligence Principles World Wide, from different research institutes, non-profit organizations, non-governmental organizations, companies, etc. The efforts aim at understanding in which degree do these different AI Principles proposals share common values, differ and complete each other. Read More
U.S. and China Go Their Own Ways With AI
As the U.S. and China appear headed for a digital cold war, competing policy approaches to the same technologies are emerging. Artificial intelligence is a prime example: Policy makers in democratic societies should, in theory, be making sure it isn’t used to promote intellectual conformity or to persecute minorities and dissidents.
The idea that AI should be ethical and benefit society has led to the emergence of multiple versions of basic principles, drafted by governments, academics and industry groups. Last year, Chinese researchers Yi Zeng, Enmeng Lu and Cunqing Huangfu identified 27 such codes and made a website on which they can be compared. It makes a somewhat eerie impression, as if the various codes form a data set on which an AI algorithm could be trained to spew forth ethical principles for its peers. Read More