I’m sort of staggered by the parallels between poetry and programming. Both are (relatively) compact modes of expression and are designed to do something in the world. They concern themselves with syntax, idiom, style, form, and clarity. Both are, in my opinion, read much less frequently and widely than they should be. I believe that, fundamentally, poems are programs—they’re just programs meant to elicit emotional responses in humans rather than perform computations on machines.
One of the ways I grew most as a poet was by doing translations, usually from French and German (I particularly like translating Jacques Prévert and Rilke). Translation makes us carefully and deeply examine our assumptions about language and how we view the world, and it’s one of the few ways we grow both as practitioners of a new skill (the foreign language) and those we take for granted (thinking and writing clearly). I’ve come to view the translation of ideas as fundamental to the critical examination of those ideas, so I do my best to apply translation as a clarifying tool whenever possible.
I became very interested in machine learning and neural networks last year. I’m in the process of reading a bunch of papers and taking a course or two on Udacity, and while doing all this I stumbled across this tutorial explaining simple neural networks and backpropagation. I found this post helpful and wonderfully readable, but even though I’m literate in Python, I found myself constantly looking up what the NumPy functions were doing. I figured that since translation has served me so well in poetry, it might be a nice exercise to translate this bit of Python into something I know a little better. So! I translated Part I of the tutorial (originally in Python + NumPy) to Clojure + the core.matrix library. It looks like this:
(ns neural.net
{:doc "A little neural network written in Clojure.
See: http://iamtrask.github.io/2015/07/12/basic-python-network/"
:author "Eric Weinstein <eric.q.weinstein@gmail.com>"}
(:require [clojure.core.matrix :as m]
[clojure.pprint :refer [pprint]]))
(defn- sigmoid
"Our nonlinear activation function. See:
https://en.wikipedia.org/wiki/Sigmoid_function"
[n]
(/ 1 (+ 1 (Math/exp (* -1 n)))))
(defn- sigmoid-deriv
"Derivative of the sigmoid function. See:
http://www.ai.mit.edu/courses/6.892/lecture8-html/sld015.htm"
[n]
(* n (- 1 n)))
(def input
"Input data!"
[[0 0 1]
[0 1 1]
[1 0 1]
[1 1 1]])
(def output
"Output data!"
[[0] [0] [1] [1]])
(defn- synapse
"Creates a synapse with random weights. Example:
(synapse 3 2)
;; => [[-0.5 -0.1]
[-0.25 -0.2]
[-0.1 -0.3]]"
[row col]
(partition col
(mapv #(* 2 (- % 0.5))
(take (* col row)
(repeatedly rand)))))
;; Create a synapse (a connection between
;; our input and hidden layers, which are
;; layer-0 and layer-1 below, respectively).
(def synapse-0 (atom (synapse 3 1)))
(defn- train
"Trains our neural network. Takes a number, n,
which is the number of iterations to run."
[n]
(dotimes [i n]
(let [layer-0 input
;; Feed forward through our hidden layer
layer-1 (mapv #(mapv sigmoid %) (m/dot layer-0 @synapse-0))
;; How much were we off by?
error (m/sub output layer-1)
;; Multiply our error by the slope of the
;; sigmoid function at the values in layer-1
delta (m/dot error (mapv #(map sigmoid-deriv %) layer-1))]
;; Update our weights
(swap! synapse-0 (partial m/add (m/dot (m/transpose layer-0) delta)))
;; How are we doing?
(pprint layer-1))))
(defn -main
[& args]
(train 10000))
This is a reasonably faithful translation from the original Python, though there are a few bits and pieces that aren’t perfect:
- I’m not jazzed about the continual rebinding (
let
insidedotimes
). - This version prints much more frequently than the original (every line rather than just at the end), but I wanted to watch the hidden layer change with each iteration.
- I’m not thrilled that I’m not seeding
rand
each time, so my results aren’t reproducible. You can do it by redefiningclojure.core/rand
:
(def r (java.util.Random. 0))
(defn rand
([] (.nextDouble r))
([n] (.nextInt r n)))
…but I’m not sure I recommend it. Redefining core functions can lead to wonky behavior, and I’m sure there’s a better way to ensure reproducibility (either a more elegant Java interop or a numerical/machine learning library that abstracts this away). The point though, I think, is that I understand the code and the library calls in a way that I wouldn’t have had I simply typed the program from memory or just read through the NumPy documentation.
Translation, like teaching, requires a thorough understanding—a filling-in of all the blank spots on the map—and I think this is crucial to deep learning (pun super intended) for everything from nineteenth-century German poetry to twenty-first century neural networks.