This is so great! Frankly, I believe that this kind of low-parameter-count high complexity optimization task is the least suitable kind of task for SGD. Bad local optima everywhere. But I didn't let this opinion of mine spoil the fun:
I changed Chamfer distance to unbiased Sinkhorn divergence (via GeomLoss), bumped arity to 4, moved randomness out of the training loop (with the goal of making training more stable), and added a LR scheduler.
Here's my notebook: https://colab.research.google.com/drive/154ffvEWpD7tTW_AIqTD...
This tree parameter set is quite nice and interpretable: https://users.renyi.hu/~daniel/tmp/ifs-christmas-tree-arity-...
Hey, I remember you from the Scratch forums back in the early '10s. It's great to see that you're still producing excellent content!
That's both amazing and beautiful. Well done!
Nice!
Upvoted for the pure passion that just oozes through the sentences. Didn’t understand half of it but that did not detract from the reading fun! :)