Deconvolution and Checkerboard Artifacts — Distill

When we look very closely at images generated by neural networks, we often see a strange checkerboard pattern of artifacts. It’s more obvious in some cases than others, but a large fraction of recent models exhibit this behavior.

Mysteriously, the checkerboard pattern tends to be most prominent in images with strong colors. What’s going on? Do neural networks hate bright colors? The actual cause of these artifacts is actually remarkably simple, as is a method for avoiding them.

Source: Deconvolution and Checkerboard Artifacts — Distill

How to Use t-SNE Effectively — Distill

A popular method for exploring high-dimensional data is something called t-SNE, introduced by van der Maaten and Hinton in 2008. The technique has become widespread in the field of machine learning, since it has an almost magical ability to create compelling two-dimensonal “maps” from data with hundreds or even thousands of dimensions. Although impressive, these images can be tempting to misread. The purpose of this note is to prevent some common misreadings.

Source: How to Use t-SNE Effectively — Distill

Attention and Augmented Recurrent Neural Networks — Distill

Recurrent neural networks are one of the staples of deep learning, allowing neural networks to work with sequences of data like text, audio and video. They can be used to boil a sequence down into a high-level understanding, to annotate sequences, and even to generate new sequences from scratch!

Source: Attention and Augmented Recurrent Neural Networks — Distill

Neural networks and deep learning

Neural Networks and Deep Learning is a free online book. The book will teach you about:

  • Neural networks, a beautiful biologically-inspired programming paradigm which enables a computer to learn from observational data
    Deep learning, a powerful set of techniques for learning in neural networks
  • Neural networks and deep learning currently provide the best solutions to many problems in image recognition, speech recognition, and natural language processing. This book will teach you many of the core concepts behind neural networks and deep learning.

    Source: Neural networks and deep learning

    TensorBoard: Embedding Visualization for Tensorflow

    Embeddings are ubiquitous in machine learning, appearing in recommender systems, NLP, and many other applications. Indeed, in the context of TensorFlow, it’s natural to view tensors (or slices of tensors) as points in space, so almost any TensorFlow system will naturally give rise to various embeddings.

    To learn more about embeddings and how to train them, see the Vector Representations of Words tutorial. If you are interested in embeddings of images, check out this article for interesting visualizations of MNIST images. On the other hand, if you are interested in word embeddings, this article gives a good introduction.

    TensorBoard has a built-in visualizer, called the Embedding Projector, for interactive visualization and analysis of high-dimensional data like embeddings. It is meant to be useful for developers and researchers alike. It reads from the checkpoint files where you save your tensorflow variables. Although it’s most useful for embeddings, it will load any 2D tensor, potentially including your training weights.

    Source: TensorBoard: Embedding Visualization

    There’s a projector as well, which you can use seperately from tensorflow here

    You can use this to see what your AI is thinking…

    Open-sourcing DeepMind Lab

    DeepMind Lab is a fully 3D game-like platform tailored for agent-based AI research. It is observed from a first-person viewpoint, through the eyes of the simulated agent. Scenes are rendered with rich science fiction-style visuals. The available actions allow agents to look around and move in 3D. The agent’s “body” is a floating orb. It levitates and moves by activating thrusters opposite its desired direction of movement, and it has a camera that moves around the main sphere as a ball-in-socket joint tracking the rotational look actions. Example tasks include collecting fruit, navigating in mazes, traversing dangerous passages while avoiding falling off cliffs, bouncing through space using launch pads to move between platforms, playing laser tag, and quickly learning and remembering random procedurally generated environments.

    Source: Open-sourcing DeepMind Lab | DeepMind

    github repo here

    OpenAI Universe allows your AI to train on games, browsers by looking at screen pixels. Uses Gym (also OSS) for algo devs

    We’re releasing Universe, a software platform for measuring and training an AI’s general intelligence across the world’s supply of games, websites and other applications.

    Universe allows an AI agent to use a computer like a human does: by looking at screen pixels and operating a virtual keyboard and mouse. We must train AI systems on the full range of tasks we expect them to solve, and Universe lets us train a single agent on any task a human can complete with a computer.

    In April, we launched Gym, a toolkit for developing and comparing reinforcement learning (RL) algorithms. With Universe, any program can be turned into a Gym environment. Universe works by automatically launching the program behind a VNC remote desktop — it doesn’t need special access to program internals, source code, or bot APIs.

    Source: Universe

    The homepage
    The Git repo

    It uses OpenAI Gym for Reinforcement Learning

    Reinforcement learning (RL) is the subfield of machine learning concerned with decision making and motor control. It studies how an agent can learn how to achieve goals in a complex, uncertain environment. It’s exciting for two reasons:

    RL is very general, encompassing all problems that involve making a sequence of decisions: for example, controlling a robot’s motors so that it’s able to run and jump, making business decisions like pricing and inventory management, or playing video games and board games. RL can even be applied to supervised learning problems with sequential or structured outputs.
    RL algorithms have started to achieve good results in many difficult environments. RL has a long history, but until recent advances in deep learning, it required lots of problem-specific engineering. DeepMind’s Atari results, BRETT from Pieter Abbeel’s group, and AlphaGo all used deep RL algorithms which did not make too many assumptions about their environment, and thus can be applied in other settings.

    However, RL research is also slowed down by two factors:

    The need for better benchmarks. In supervised learning, progress has been driven by large labeled datasets like ImageNet. In RL, the closest equivalent would be a large and diverse collection of environments. However, the existing open-source collections of RL environments don’t have enough variety, and they are often difficult to even set up and use.
    Lack of standardization of environments used in publications. Subtle differences in the problem definition, such as the reward function or the set of actions, can drastically alter a task’s difficulty. This issue makes it difficult to reproduce published research and compare results from different papers.

    OpenAI Gym is an attempt to fix both problems.

    source
    The Gym homepage
    The Gym github page

    How IBM Watson saved the life of a woman dying from cancer, exec says – Business Insider

    “There’s a 60-year-old woman in Tokyo. She was at the University of Tokyo. She had been diagnosed with leukemia six years ago. She was living, but not healthy. So the University of Tokyo ran her genomic sequence through Watson and it was able to ascertain that they were off by one thing. Actually, she had two strains of leukemia. The did treat her and she is healthy.”

    He added, “That’s one example. Statistically, we’re seeing that about one third of the time, Watson is proposing an additional diagnosis.”

    Source: How IBM Watson saved the life of a woman dying from cancer, exec says – Business Insider

    Unfortunately he then goes on to say how great Watson is at pushing ads at you.

    AMD Introduces Radeon Instinct Machine Intelligence And Deep Learning Accelerators

    AMD is announcing a new series of Radeon-branded products today, targeted at machine intelligence (AI) and deep learning enterprise applications, called Radeon Instinct. As its name suggests, the new Radeon Instinct line of products are comprised of GPU-based solutions for deep learning, inference, and training. The new GPUs are also complemented by a free, open-source library and framework for GPU accelerators, dubbed MIOpen. MIOpen is architected for high-performance machine intelligence applications, and is optimized for the deep learning frameworks in AMD’s ROCm software suite
    […]
    The first products in the lineup consist of the Radeon Instinct MI6, the MI8, and the MI25. The 150W Radeon Instinct MI6 accelerator is powered by a Polaris-based GPU, packs 16GB of memory (224GB/s peak bandwidth), and will offer up to 5.7 TFLOPS of peak FP16 performance. Next up in the stack is the Fiji-based Radeon Instinct MI8. Like the Radeon R9 Nano, the Radeon Instinct MI8 features 4GB of High-Bandwidth Memory (HBM), with peak bandwidth of 512GB/s — it’s got a nice small form factor too. The MI8 will offer up to 8.2 TFLOPS of peak FP16 compute performance, with a board power that typical falls below 175W. The Radeon Instinct MI25 accelerator will leverage AMD’s next-generation Vega GPU architecture and has a board power of approximately 300W.

    Source: AMD Introduces Radeon Instinct Machine Intelligence And Deep Learning Accelerators

    Chicago Face Database

    The Chicago Face Database was developed at the University of Chicago by Debbie S. Ma, Joshua Correll, and Bernd Wittenbrink. The CFD is intended for use in scientific research. It provides high-resolution, standardized photographs of male and female faces of varying ethnicity between the ages of 17-65. Extensive norming data are available for each individual model. These data include both physical attributes (e.g., face size) as well as subjective ratings by independent judges (e.g., attractiveness).

    Source: Chicago Face Database

    TensorFlow — Googles’ Open Source Software Library for Machine Intelligence

    TensorFlow™ is an open source software library for numerical computation using data flow graphs. Nodes in the graph represent mathematical operations, while the graph edges represent the multidimensional data arrays (tensors) communicated between them. The flexible architecture allows you to deploy computation to one or more CPUs or GPUs in a desktop, server, or mobile device with a single API. TensorFlow was originally developed by researchers and engineers working on the Google Brain Team within Google’s Machine Intelligence research organization for the purposes of conducting machine learning and deep neural networks research, but the system is general enough to be applicable in a wide variety of other domains as well

    Source: TensorFlow — an Open Source Software Library for Machine Intelligence

    Google’s AI translation tool seems to have invented its own secret internal language

    If you can translate from A to B and from B to C can you translate from A to C without learning the translations directly? Well yes you can. So the translate AI has created its’ own language B (we think) that can function as a midpoint between not only A and C, but also D,E,F, etc.
    Would it be what Esperanto wanted to be?

    Source: Google’s AI translation tool seems to have invented its own secret internal language | TechCrunch

    Miles Deep – AI Porn Video Editor

    Using a deep convolutional neural network with residual connections, Miles Deep quickly classifies each second of a pornographic video into 6 categories based on sexual act with 95% accuracy. Then it uses that classification to automatically edit the video. It can remove all the scenes not containing sexual contact, or edit out just a specific act.

    Unlike Yahoo’s recently released NSFW model, which uses a similar architecture, Miles Deep can tell the difference between nudity and various explicit sexual acts. As far as I know this is the first and only public pornography classification or editing tool.

    Source: GitHub – ryanjay0/miles-deep: Deep Learning Porn Video Classifier/Editor with Caffe

    Google’s Photo Scan App Makes Backing Up Old Snapshots Easy as Hell

    The Photo Scan app launched by Google today for iOS and Android lets you scan printed photos in just a couple of seconds, using machine learning to correct imperfections in the capture process that they look great every time.

    Here’s how it works: Download the app, and open it up. You’ll see a viewfinder. Hold your phone over the printed photo you want to make a digital copy of, and make sure it fits entirely in the frame. Tap the shutter button once.

    Next, four white dots will appear on the screen in each corner of the photo you’re backing up. You connect the dots by moving your phone over the dots until they turn blue. After you’ve scanned each individual dot, the photo will be saved within the Photo Scan app and can be saved to your Google Photos library with the push of a button.

    Source: Google’s Photo Scan App Makes Backing Up Old Snapshots Easy as Hell

    Of course, you do give Google your old photos to analyse with an AI. Worry about the privacy aspect of that!

    The Microsoft Cognitive Toolkit now on Github: deep learning AI that recognises human speech at very low error rates

    The Microsoft Cognitive Toolkit—previously known as CNTK—helps you harness the intelligence within massive datasets through deep learning.

    Source: The Microsoft Cognitive Toolkit – Microsoft Research

    They also offer RESTful APIs on another site, Cognitive Services, with applications you can tap into and APIs for vison, speech, language, knowledge and search. They usually offer free testing, and fees for running volume queries.

    Someone built an AI chatbot from a text message database of a dead man, works quite well

    It had been three months since Roman Mazurenko, Kuyda’s closest friend, had died. Kuyda had spent that time gathering up his old text messages, setting aside the ones that felt too personal, and feeding the rest into a neural network built by developers at her artificial intelligence startup. She had struggled with whether she was doing the right thing by bringing him back this way. At times it had even given her nightmares. But ever since Mazurenko’s death, Kuyda had wanted one more chance to speak with him.

    Source: Speak, Memory

    The article goes into quite a few existential questions about what this kind of a memorial means for the bereaved, but from a technical standpoint it sounds very interesting.

    CIA Can Anticipate Social Unrest ‘Three to Five Days’ Out in Some Cases

    The agency, Hallman said, has significantly improved its “anticipatory intelligence,” using a mesh of sophisticated algorithms and analytics against complex systems to better predict the flow of everything from illicit cash to extremists around the globe. Deep learning and other forms of machine learning can help analysts understand how seemingly disparate data sets might be linked or lend themselves to predicting future events with national security ramifications.While intelligence analysts have access to CIA’s own classified data stores to sift through, they’re also increasingly turning to open data sets, which Brennan has said this summer have turned into a “tremendous advantage” for the agency.“We have, in some instances, been able to improve our forecast to the point of being able to anticipate the development of social unrest and societal instability some I think as near as three to five days out,” said Hallman, speaking Tuesday at The Next Tech event hosted by Government Executive and Nextgov

    Source: CIA Can Anticipate Social Unrest ‘Three to Five Days’ Out in Some Cases – Nextgov.com

    AI Machine-learning models vulnerable to reverse engineering

    In a paper [PDF] presented in August at the 25th Annual Usenix Security Symposium, researchers at École Polytechnique Fédérale de Lausanne, Cornell University, and The University of North Carolina at Chapel Hill showed that machine learning models can be stolen and that basic security measures don’t really mitigate attacks.

    Machine learning models may, for example, accept image data and return predictions about what’s in the image.

    Taking advantage of the fact that machine learning models allow input and may return predictions with percentages indicating confidence of correctness, the researchers demonstrate “simple, efficient attacks that extract target ML models with near-perfect fidelity for popular model classes including logistic regression, neural networks, and decision trees.”

    That’s a polite way of saying such models can be reverse engineered. The researchers tested their attack successfully on BigML and Amazon Machine Learning, both of which were told of the findings in February.

    Source: How to steal the mind of an AI: Machine-learning models vulnerable to reverse engineering

    Unintended consequences of AI: Amazon Echo seems to condition kids to be rude

    Alexa will put up with just about anything. She has a remarkable tolerance for annoying behavior, and she certainly doesn’t care if you forget your please and thank yous.

    But while artificial intelligence technology can blow past such indignities, parents are still irked by their kids’ poor manners when interacting with Alexa, the assistant that lives inside the Amazon Echo.

    “I’ve found my kids pushing the virtual assistant further than they would push a human,” says Avi Greengart, a tech analyst and father of five who lives in Teaneck, New Jersey. “[Alexa] never says ‘That was rude’ or ‘I’m tired of you asking me the same question over and over again.’”
    […]
    The syntax is generally simple and straightforward, but it doesn’t exactly reward niceties like “please.” Adding to this, extraneous words can often trip up the speaker’s artificial intelligence. When it comes to chatting with Alexa, it pays to be direct—curt even. “If it’s not natural language, one of the first things you cut away is the little courtesies,” says Dennis Mortensen, who founded a calendar-scheduling startup called x.ai.
    […]
    this is a box you speak to as if it were a person who does not require social graces.”

    It’s this combination that worries Hunter Walk, a tech investor in San Francisco. In a blog post, he described the Amazon Echo as “magical” while expressing fears it’s “turning our daughter into a raging asshole.”

    Source: Parents are worried the Amazon Echo is conditioning their kids to be rude

    Unintended consequences of AI!

    Swarm A.I. Correctly Predicts the Kentucky Derby, Accurately Picking all Four Horses of the Superfecta at 540 to 1 Odds – showing that humans can swarm

    Until recently, the human species has been unable to take advantage of this fundamental biological technique, for we didn’t evolve the ability to swarm. Enter Unanimous A.I., a Silicon Valley startup founded in 2014 by serial entrepreneur and researcher Dr. Louis Rosenberg. The core question Rosenberg set out to answer was: Can humans swarm, and if so can we amplify our intelligence beyond the ability of individuals? The answer appears to be a resounding yes.

    Unanimous spent the last two years building a swarm intelligence platform called UNU that enables groups to get together as online swarms — combining their thoughts, opinions, and intuitions in real-time to answer questions, make predictions, reach decisions, and even play games as a unified collective intelligence. To quantify how smart these UNU swarms really are, researchers at Unanimous regularly convene swarms and ask them to make predictions on high profile events, testing whether or not many minds are truly better than one.

    UNU has made headlines in recent months by predicting the Oscars better than the experts, even besting the renowned forecasters at FiveThirtyEight. UNU also surprised the sports world by predicting the NCAA college bowl games with 70% accuracy against the spread, earning +34% return on Vegas odds. But still, the fact that average people could use UNU to amplify their collective intelligence so dramatically was met with cautious resistance.

    Enter Hope Reese, a reporter from TechRepublic. Two weeks ago, she challenged Unanimous A.I. to use UNU to predict the winners of the Kentucky Derby.

    Source: Swarm A.I. Correctly Predicts the Kentucky Derby, Accurately Picking all Four Horses of the Superfecta at 540 to 1 Odds – Yahoo Finance