Implementing a self-organizing map, part 2: Nodes and the map

2025-03-11

With Vec and Hex in place, we're now ready to implement the SOM node.

struct Node { const Hex hex; Vec weight; Node(const Hex& hex) : hex(hex), weight() {} Node(const Node&) = default; };

Before we can construct a SOM we'll also need something to generate the map grid. I have a function here that generates a rectangular grid of Hexs, but you could replace this with something that generates a funkier shape.

std::vector<Hex> GenerateHexGrid(int rows_top, int rows_bottom, int cols_left, int cols_right) { std::vector<Hex> grid; for (int q = cols_left; q <= cols_right; ++q) { int q_offset = q >> 1; for (int r = rows_top - q_offset; r <= rows_bottom - q_offset; ++r) { grid.push_back(Hex(q, r, -q - r)); } } return grid; }

The SOM

Now we can get started on the meat of this exercise: implementing the SOM itself. First, let's create the Som class with a constructor, a destructor and a getter for the vector that holds the nodes.

class Som { public: using NodesT = std::vector<Node*>; static constexpr unsigned kTrainingEpochLimit = 10; private: NodesT nodes_; unsigned current_training_epoch_; public: Som(int rows, int cols) { int rows_bottom = rows / 2; int rows_top = -(rows - rows_bottom - 1); int cols_right = cols / 2; int cols_left = -(cols - cols_right - 1); auto hexes = GenerateHexGrid(rows_top, rows_bottom, cols_left, cols_right); for (auto hex : hexes) { nodes_.push_back(new Node(hex)); nodes_.back()->weight.Randomize(0.0, 255.0); } current_training_epoch_ = 0; } ~Som() { for (auto node : nodes_) { delete node; } } const NodesT& GetNodes() const { return nodes_; } ...

The SOM constructor accepts the numbers of rows and columns that the grid should contain. Since we will be using the weights of the nodes to define their color when they're drawn and we want the initial map to be made up of random colors, the components of the weights are drawn randomly from the range [0.0, 255.0). We could just as easily use any other range, such as [0.0, 1.0) for instance, which would make the initial map black.

We also set a training epoch limit, which limits the number of epochs that the map can be trained to 10. This is very low, but for learning colors it should be enough. The training epoch limit is used in calculating the learning rate.

Now we get to where the magic happens, the training method.

class Som { ... public: bool TrainOneEpoch(const std::vector<Vec>& input_batch) { ++current_training_epoch_; if (current_training_epoch_ > kTrainingEpochLimit) { return false; } double learning_rate = 1 - (static_cast<double>(current_training_epoch_) / static_cast<double>(kTrainingEpochLimit)); for (auto input_vec : input_batch) { Node* bmu = FindBmu(input_vec); assert(bmu); UpdateWeights(bmu, input_vec, learning_rate); } return true; } ... private: Node* FindBmu(const Vec& input_vec) const { Node* bmu = nullptr; for (auto& node : nodes_) { if ((bmu == nullptr) || (EuclideanDistance(input_vec, node->weight) < EuclideanDistance(input_vec, bmu->weight))) { bmu = node; } } return bmu; } void UpdateWeights(Node* bmu, const Vec& input_vec, double learning_rate) { for (auto& node : nodes_) { double neighborhood_coeff = 1 / std::pow(2, HexDistance(bmu->hex, node->hex)); node->weight += neighborhood_coeff * learning_rate * (input_vec - node->weight); } } };

TrainOneEpoch() takes a batch of training data vectors as input and uses them to train the map. The learning rate is calculated based on the current training epoch and the training epoch limit such that it decreases as the training progresses, which means that node weights are adjusted more drastically in the beginning. The BMU is sought for each training data sample, after which both are used to update the weights of all the nodes in the map. The neighborhood function used here is one that applies the update to the BMU in full and to other nodes based on their distance to the BMU such that each unit increase in the distance halves the update amount from the previous. A good alternative to this would be the Gaussian.

With all of this in place we now have a functioning SOM. Now all that's left is to build something around it that will allow us to observe how it works, which we'll do in part 3.