\

Understanding Neural Network, Visually

189 points - last Tuesday at 2:49 PM

Source
  • tpdly

    today at 5:31 PM

    Lovely visualization. I like the very concrete depiction of middle layers "recognizing features", that make the whole machine feel more plausible. I'm also a fan of visualizing things, but I think its important to appreciate that some things (like 10,000 dimension vector as the input, or even a 100 dimension vector as an output) can't be concretely visualized, and you have to develop intuitions in more roundabout ways.

    I hope make more of these, I'd love to see a transformer presented more clearly.

    • helloplanets

      today at 5:15 PM

      For the visual learners, here's a classic intro to how LLMs work: https://bbycroft.net/llm

      • esafak

        today at 4:25 PM

        This is just scratching the surface -- where neural networks were thirty years ago: https://en.wikipedia.org/wiki/MNIST_database

        If you want to understand neural networks, keep going.

        • shrekmas

          today at 10:22 PM

          As someone who does not use Twitter, I suggest adding RSS to your site.

          • brudgers

            yesterday at 5:30 AM

            The original Show HN, https://news.ycombinator.com/item?id=44633725

            • 8cvor6j844qw_d6

              today at 7:36 PM

              Oh wow, this looks like a 3d render of a perceptron when I started reading about neural networks. I guess essentially neural networks are built based on that idea? Inputs > weight function to to adjust the final output to desired values?

                • adammarples

                  today at 10:05 PM

                  Yes, vanilla neural networks are just lots of perceptrons

              • jazzpush2

                today at 7:48 PM

                I love this visual article as well:

                https://mlu-explain.github.io/neural-networks/

                • ge96

                  today at 5:36 PM

                  I like the style of the site it has a "vintage" look

                  Don't think it's moire effect but yeah looking at the pattern

                • jetfire_1711

                  today at 8:56 PM

                  Spent 10 minutes on the site and I think this is where I'll start my day from next week! I just love visual based learning.

                  • cwt137

                    today at 6:08 PM

                    This visualizations reminds me of the 3blue1brown videos.

                      • giancarlostoro

                        today at 6:12 PM

                        I was thinking the same thing. Its at least the same description.

                    • today at 8:52 PM

                      • 4fterd4rk

                        today at 3:45 PM

                        Great explanation, but the last question is quite simple. You determine the weights via brute force. Simply running a large amount of data where you have the input as well as the correct output (handwriting to text in this case).

                          • ggambetta

                            today at 4:35 PM

                            "Brute force" would be trying random weights and keeping the best performing model. Backpropagation is compute-intensive but I wouldn't call it "brute force".

                              • Ygg2

                                today at 4:58 PM

                                "Brute force" here is about the amount of data you're ingesting. It's no Alpha Zero, that will learn from scratch.

                                  • jazzpush2

                                    today at 7:49 PM

                                    What? Either option requires sufficient data. Brute force implies iterating over all combinations until you find the best weights. Back-prop is an optimization technique.

                        • artemonster

                          today at 7:15 PM

                          I get 3fps on my chrome, most likely due to disabled HW acceleration

                            • nerdsniper

                              today at 7:27 PM

                              High FPS on Safari M2 MBP.

                          • anon291

                            today at 6:53 PM

                            Nice visuals, but misses the mark. Neural networks transform vector spaces, and collect points into bins. This visualization shows the structure of the computation. This is akin to displaying a Matrix vector multiplication in Wx + b notation, except W,x,and b have more exciting displays.

                            It completely misses the mark on what it means to 'weight' (linearly transform), bias (affine transform) and then non-linearly transform (i.e, 'collect') points into bins

                              • titzer

                                today at 7:28 PM

                                > but misses the mark

                                It doesn't match the pictures in your head, but it nevertheless does present a mental representation the author (and presumably some readers) find useful.

                                Instead of nitpicking, perhaps pointing to a better visualization (like maybe this video: https://www.youtube.com/watch?v=ChfEO8l-fas) could help others learn. Otherwise it's just frustrating to read comments like this.

                            • pks016

                              today at 6:37 PM

                              Great visualization!

                              • javaskrrt

                                today at 5:45 PM

                                very cool stuff