\

Show HN: I built a tiny LLM to demystify how language models work

806 points - today at 12:20 AM


Built a ~9M param LLM from scratch to understand how they actually work. Vanilla transformer, 60K synthetic conversations, ~130 lines of PyTorch. Trains in 5 min on a free Colab T4. The fish thinks the meaning of life is food.

Fork it and swap the personality for your own character.

Source
  • thomasfl

    today at 1:42 PM

    Is there some documentation for this? The code is probably the simplest (Not So) Large Language Model implementation possible, but it is not straight forward to understand for developers not familiar with multi-head attention, ReLU FFN, LayerNorm and learned positional embeddings.

    This projects shares similarities with Minix. Minix is still used at universities as an educational tool for teaching operating system design. Minix is the operating system that taught Linus Torvalds how to design (monolithic) operating systems. Similarly having students adding capabilities to GuppyLM is a good way to learn LLM design.

      • achenatx

        today at 1:58 PM

        give the code to an LLM and have a discussion about it.

          • dominotw

            today at 3:45 PM

            does this work? there is no more need for writing high level docs?

              • arcanemachiner

                today at 4:09 PM

                > does this work?

                Absolutely. If you loaded this into an agentic coding harness with a decent model, I can practically guarantee it would be able to help you figure out what's going on.

                > there is no more need for writing high level docs?

                Absolutely not. That would be like exploring a cave without a flashlight, knowing that you could just feel your way around in the dark instead.

                Code is not always self-documenting, and can often tell you how it was written, but not why.

                  • stronglikedan

                    today at 5:35 PM

                    > If you loaded this into an agentic coding harness with a decent model, I can practically guarantee it would be able to help you figure out what's going on.

                    My non-coder but technically savvy boss has been doing this lately to great success. It's nice because I spend less time on it since the model has taken my place for the most part.

                      • libria

                        today at 7:35 PM

                        > since the model has taken my place for the most part

                        Hah, you realize the same thing is going on in your boss's head right? The pie chart of Things-I-Need-stronglikedan-For just shrank tiny bit...

                • sigmoid10

                  today at 4:07 PM

                  There are so many blogs and tutorials about this stuff in particular, I wouldn't worry about it being outside the training data distribution for modern LLMs. If you have a scarce topic in some obscure language I'd be more careful when learning from LLMs.

                  • bigmadshoe

                    today at 4:04 PM

                    LLMs can tell you what the code does but not why the developer chose to do it that way.

                    Also, large codebases are harder to understand. But projects like these are simple to discuss with an LLM.

                      • stronglikedan

                        today at 5:37 PM

                        > LLMs can tell you what the code does but not why the developer chose to do it that way.

                        Do LLMs not take comments into consideration? (Serious question - I'm just getting into this stuff)

                          • dr_hooo

                            today at 6:56 PM

                            They do (it's just text), if they are there...

        • fg137

          today at 10:15 AM

          How does this compare to Andrej Karpathy's microgpt (https://karpathy.github.io/2026/02/12/microgpt/) or minGPT (https://github.com/karpathy/minGPT)?

            • armanified

              today at 11:37 AM

              I haven't compared it with anything yet. Thanks for the suggestion; I'll look into these.

              • BrokenCogs

                today at 1:13 PM

                Who cares how it compares, it's not a product it's a cool project

                  • tantalor

                    today at 1:28 PM

                    Even cool projects can learn from others. Maybe they missed something that could benefit the project, or made some interesting technical choice that gives a different result.

                    For the readers/learners, it's useful to understand the differences so we know what details matter, and which are just stylistic choices.

                    This isn't art; it's science & engineering.

                      • BrokenCogs

                        today at 1:52 PM

                        But it isn't the OP's responsibility to compare their project to all other projects. The GP could themselves perform the comparison and post their thoughts instead of asking an open ended question.

                          • philipallstar

                            today at 2:23 PM

                            > it isn't the OP's responsibility to compare their project to all other projects

                            No one, including the GP, said it was.

                            • fg137

                              today at 2:54 PM

                              It isn't, but such information will be immensely helpful to anyone who wants to learn from such projects. Some tutorials are objectively better than others, and learners can benefit from such information.

                              • tantalor

                                today at 2:02 PM

                                100% agree, I didn't mean to imply that OP is responsible for that, or that the (lack of) comparison detracts in any way from the work.

                        • stronglikedan

                          today at 5:39 PM

                          > Who cares how it compares

                          Well, the person who asked the question, for one. I'm sure they're not the only one. Best not to assume why people are asking though, so you can save time by not writing irrelevant comments.

                          • layer8

                            today at 4:07 PM

                            Microgpt isn’t a product either. Are you saying that differences between cool projects aren’t worth thinking and conversing about?

                            • today at 1:45 PM

                      • totetsu

                        today at 9:12 AM

                        https://bbycroft.net/llm has 3d Visualization of tiny example LLM layers that do a very good job at showing what is going on (https://news.ycombinator.com/item?id=38505211)

                          • armanified

                            today at 11:53 AM

                            Pretty neat! I'll definitely take a deeper look into this.

                            • maverickxone

                              today at 11:11 AM

                              have little to do with this, but i have to say your project are indeed pretty cool! Consider adding some more UI?

                              • skramzy

                                today at 1:22 PM

                                Neat!

                            • ordinarily

                              today at 2:57 AM

                              It's genuinely a great introduction to LLMs. I built my own awhile ago based off Milton's Paradise Lost: https://www.wvrk.org/works/milton

                              • algoth1

                                today at 11:54 AM

                                This really makes me think if it would be feasible to make an llm trained exclusively on toki pona (https://en.wikipedia.org/wiki/Toki_Pona)

                                  • MarkusQ

                                    today at 2:46 PM

                                    There isn't enough training data though, is there? The "secret sauce" of LLMs is the vast amount of training data available + the compute to process it all.

                                      • algoth1

                                        today at 4:51 PM

                                        I think you could probably feed a copy of a toki pona grammar book to a big model, and have it produce ā€˜infinite’ training data

                                          • MarkusQ

                                            today at 8:08 PM

                                            This is essentially a distillation on the bigger model; you'd wind up surfacing a lot of artifacts from the host model, amplifying them in the same way repeated photocopying introduces errors.

                                            https://dailyai.com/2025/05/create-a-replica-of-this-image-d...

                                            • eden-u4

                                              today at 6:34 PM

                                              There are not enough samples in that book to generate new "infinite" data.

                                  • neurworlds

                                    today at 1:16 PM

                                    Cool project. I'm working on something where multiple LLM agents share a world and interact with each other autonomously. One thing that surprised me is how much the "world" matters — same model, same prompt, but put it in a system with resource constraints, other agents, and persistent memory, the behavior changes dramatically. Made me realize we spend too much time optimizing the model and not enough thinking about the environment it operates in.

                                    • mudkipdev

                                      today at 5:55 AM

                                      This is probably a consequence of the training data being fully lowercase:

                                      You> hello Guppy> hi. did you bring micro pellets.

                                      You> HELLO Guppy> i don't know what it means but it's mine.

                                        • functional_dev

                                          today at 6:51 AM

                                          Great find! It appears uppercase tokens are completely unknonw to the tokenizer.

                                          But the character still comes through in response :)

                                      • hackerman70000

                                        today at 8:17 AM

                                        Finally an LLM that's honest about its world model. "The meaning of life is food" is arguably less wrong than what you get from models 10,000x larger

                                          • amelius

                                            today at 11:19 AM

                                            It's arguably even better than the most famous answer to that question.

                                          • zkmon

                                            today at 11:53 AM

                                            Meaning/goal of life is to reproduce. Food (and everything else) is only a means to it. Reproduction is the only root goal given by nature to any life form. All resources and qualities are provided are only to help mating.

                                              • tantalor

                                                today at 1:31 PM

                                                Reproduction is the goal of genes.

                                                Food (not dying) is the goal of organisms.

                                                  • philote

                                                    today at 2:23 PM

                                                    I'd argue genes nor life has a "goal". They are what they are because they've been successful at continuing their existence. Would you say a rock's goal is not to get broken?

                                                      • tantalor

                                                        today at 2:36 PM

                                                        Only because genes/organisms can make choices (changes to its programming, or decisions) to optimize their path towards their goal.

                                                        A rock is maybe not a good counterexample, but a crystal is because it can grow over time. So in some sense, it tries not to break. However a crystal cannot make any choices; it's behavior is locked into the chemistry it starts with.

                                                • amelius

                                                  today at 12:07 PM

                                                  Then why are reproductive rates so low in western countries?

                                                  https://en.wikipedia.org/wiki/List_of_countries_by_total_fer...

                                                    • darepublic

                                                      today at 12:42 PM

                                                      The western lifestyle is an evolutionary dead end?

                                                        • vixen99

                                                          today at 1:10 PM

                                                          It seems that some in the West want it to be and are working hard to make it so.

                                                  • hca

                                                    today at 2:49 PM

                                                    No, evolution has encoded lust. It has not yet allowed for condoms. But it's a process.

                                            • BiraIgnacio

                                              today at 5:39 PM

                                              Nice work and thanks for sharing it!

                                              Now, I ask, have LLMs ben demystified to you? :D

                                              I am still impressed how much (for the most part) trivial statistics and a lot of compute can do.

                                              • rpdaiml

                                                today at 2:44 PM

                                                This is a nice idea. A tiny implementation can be way more useful for learning than yet another wrapper around a big model, especially if it keeps the training loop and inference path small enough to read end to end.

                                                • zwaps

                                                  today at 5:46 AM

                                                  I like the idea, just that the examples are reproduced from the training data set.

                                                  How does it handle unknown queries?

                                                    • armanified

                                                      today at 11:41 AM

                                                      It mostly doesn't, at 9M it has very limited capacity. The whole idea of this project is to demonstrate how Language Models work.

                                                      • today at 8:46 AM

                                                    • bblb

                                                      today at 9:01 AM

                                                      Could it be possible to train LLM only through the chat messages without any other data or input?

                                                      If Guppy doesn't know regular expressions yet, could I teach it to it just by conversation? It's a fish so it wouldn't probably understand much about my blabbing, but would be interesting to give it a try.

                                                      Or is there some hard architectural limit in the current LLM's, that the training needs to be done offline and with fairly large training set.

                                                        • roetlich

                                                          today at 11:15 AM

                                                          What does "done offline" mean? Otherwise you are limited by context window.

                                                          • tatrions

                                                            today at 1:58 PM

                                                            [flagged]

                                                        • bharat1010

                                                          today at 5:39 PM

                                                          This is such a smart way to demystify LLMs. I really like that GuppyLM makes the whole pipeline feel approachable..great work

                                                          • Leomuck

                                                            today at 2:34 PM

                                                            Wow that is such a cool idea! And honestly very much needed. LLMs seem to be this blackbox nobody understands. So I love every effort to make that whole thing less mysterious. I will definitely have a look at dabbling with this, may it not be a goldfish LLM :)

                                                            • CaseFlatline

                                                              today at 2:49 PM

                                                              I am trying to find how the synthetic data was created (looking through the repo) and didn't find it. Maybe I am missing it - Would love to see the prompts and process on that aspect of the training data generation!

                                                            • cbdevidal

                                                              today at 3:13 AM

                                                              > you're my favorite big shape. my mouth are happy when you're here.

                                                              Laughed loudly :-D

                                                                • vunderba

                                                                  today at 4:07 AM

                                                                  This is a direct output from the synthetic training data though - wonder if there is a bit of overfitting going on or it’s just a natural limitation of a much smaller model.

                                                              • jzer0cool

                                                                today at 3:04 PM

                                                                Does this work by just training once with next token prediction? Want to understand better how it creates fluent sentences if anyone can provide insights.

                                                                • EmilioOldenziel

                                                                  today at 4:31 PM

                                                                  Building it yourself is always the best test if you really understand how it works.

                                                                  • jbethune

                                                                    today at 4:15 PM

                                                                    Forked. Very cool. I appreciate the simplicity and documentation.

                                                                    • kaipereira

                                                                      today at 5:22 AM

                                                                      This is so cool! I'd love to see a write-up on how made it, and what you referenced because designing neural networks always feel like a maze ;)

                                                                      • Duplicake

                                                                        today at 10:48 AM

                                                                        I love this! Seems like it can't understand uppercase letters though

                                                                          • armanified

                                                                            today at 11:51 AM

                                                                            Uppercase letters were intentionally ignored.

                                                                        • brcmthrowaway

                                                                          today at 5:58 AM

                                                                          Why are there so many dead comments from new accounts?

                                                                            • 59nadir

                                                                              today at 9:06 AM

                                                                              Because despite what HN users seem to think, HN is a LLM-infested hellscape to the same degree as Reddit, if not more.

                                                                                • wiseowise

                                                                                  today at 10:18 AM

                                                                                  You’re absolutely right! HN isn’t just LLM-infested hellscape, it’s a completely new paradigm of machine assisted chocolate-infused information generation.

                                                                                    • toyg

                                                                                      today at 10:31 AM

                                                                                      Just let me know which type of information goo you'd like me to generate, and I'll tailor the perfect one for you.

                                                                                  • siva7

                                                                                    today at 11:55 AM

                                                                                    But what should we do? The parent company isn't transparent about communicating the seriousness of this problem

                                                                                • loveparade

                                                                                  today at 7:13 AM

                                                                                  It really seems it's mostly AI comments on this. Maybe this topic is attractive to all the bots.

                                                                                    • armanified

                                                                                      today at 12:45 PM

                                                                                      This title might have triggered something in those bots; most of them have sneaky AI SaaS links in their bio.

                                                                                      Honestly, I never expected this post to become so popular. It was just the outcome of a weekend practice session.

                                                                                  • AlecSchueler

                                                                                    today at 6:00 AM

                                                                                    They all seem to be slop comments.

                                                                                • ankitsanghi

                                                                                  today at 5:15 AM

                                                                                  Love it! I think it's important to understand how the tools we use (and will only increasingly use) work under the hood.

                                                                                  • nobodyandproud

                                                                                    today at 2:25 PM

                                                                                    Thanks. Tinkering is how I learn and this is what I’ve been looking for.

                                                                                    • drincanngao

                                                                                      today at 11:27 AM

                                                                                      I was going to suggest implementing RoPE to fix the context limit, but realized that would make it anatomically incorrect.

                                                                                        • armanified

                                                                                          today at 11:36 AM

                                                                                          I intentionally removed all optimizations to keep it vanilla.

                                                                                      • fawabc

                                                                                        today at 10:32 AM

                                                                                        how did you generate the synthetic data?

                                                                                        • winter_blue

                                                                                          today at 3:53 PM

                                                                                          This is amazing work. Thank you.

                                                                                          • amelius

                                                                                            today at 11:18 AM

                                                                                            > A 9M model can't conditionally follow instructions

                                                                                            How many parameters would you need for that?

                                                                                              • armanified

                                                                                                today at 11:50 AM

                                                                                                My initial idea was to train a navigation decision model with 25M parameters for a Raspberry Pi, which, in testing, was getting about 60% of tool calls correct. IMO, it seems like around 20M parameters would be a good size for following some narrow & basic language instructions.

                                                                                                  • amelius

                                                                                                    today at 12:39 PM

                                                                                                    Ok. This makes me wonder about a broader question. Is there a scientific approach showing a pyramid of cognitive functions, and how many parameters are (minimally) required for each layer in this pyramid?

                                                                                            • SilentM68

                                                                                              today at 2:22 AM

                                                                                              Would have been funny if it were called "DORY" due to memory recall issues of the fish vs LLMs similar recall issues :)

                                                                                                • armanified

                                                                                                  today at 11:59 AM

                                                                                                  OMG! Why didn't I thought fo this first :P

                                                                                              • kubrador

                                                                                                today at 5:33 AM

                                                                                                how's it handle longer context or does it start hallucinating after like 2 sentences? curious what the ceiling is before the 9M params

                                                                                                • gnarlouse

                                                                                                  today at 3:32 AM

                                                                                                  I... wow, you made an LLM that can actually tell jokes?

                                                                                                    • murkt

                                                                                                      today at 7:15 AM

                                                                                                      With 9M params it just repeats the joke from a training dataset.

                                                                                                  • ben8bit

                                                                                                    today at 9:09 AM

                                                                                                    This is really great! I've been wanting to do something similar for a while.

                                                                                                    • rclkrtrzckr

                                                                                                      today at 6:57 AM

                                                                                                      I could fork it and create TrumpLM. Not a big leap, I suppose.

                                                                                                        • search_facility

                                                                                                          today at 8:45 AM

                                                                                                          probably 8M params are too much even :)

                                                                                                            • danparsonson

                                                                                                              today at 10:16 AM

                                                                                                              As long as you use the best parameters then it doesn't matter

                                                                                                              • wiseowise

                                                                                                                today at 10:19 AM

                                                                                                                Grab her by the pointer.

                                                                                                        • NyxVox

                                                                                                          today at 3:47 AM

                                                                                                          Hm, I can actually try the training on my GPU. One of the things I want to try next. Maybe a bit more complex than a fish :)

                                                                                                          • rahen

                                                                                                            today at 1:36 PM

                                                                                                            I don't mean to be 'that guy', but after a quick review, this really feels like low-effort AI slop to me.

                                                                                                            There is nothing wrong using AI tools to write code, but nothing here seems to have taken more than a generic 'write me a small LLM in PyTorch' prompt, or any specific human understanding.

                                                                                                            The bar for what constitutes an engineering feat on HN seems to have shifted significantly.

                                                                                                            • ananandreas

                                                                                                              today at 10:16 AM

                                                                                                              Great and simple way to bridge the gap between LLMs and users coming in to the field!

                                                                                                              • cpldcpu

                                                                                                                today at 8:04 AM

                                                                                                                Love it! Great idea for the dataset.

                                                                                                                • monksy

                                                                                                                  today at 6:35 AM

                                                                                                                  Is this a reference from the Bobiverse?

                                                                                                                  • nullbyte808

                                                                                                                    today at 2:10 AM

                                                                                                                    Adorable! Maybe a personality that speaks in emojis?

                                                                                                                      • armanified

                                                                                                                        today at 11:57 AM

                                                                                                                        OMG! You just gave me the next idea..

                                                                                                                    • Vektorceraptor

                                                                                                                      today at 1:57 PM

                                                                                                                      Haha, funny name :)

                                                                                                                      • AndrewKemendo

                                                                                                                        today at 1:53 AM

                                                                                                                        I love these kinds of educational implementations.

                                                                                                                        I want to really praise the (unintentional?) nod to Nagel, by limiting capabilities to representation of a fish, the user is immediately able to understand the constraints. It can only talk like a fish cause it’s very simple

                                                                                                                        Especially compared to public models, thats a really simple correspondence to grok intuitively (small LLM > only as verbose as a fish, larger LLM > more verbose) so kudos to the author for making that simple and fun.

                                                                                                                          • dvt

                                                                                                                            today at 2:25 AM

                                                                                                                            > the user is immediately able to understand the constraints

                                                                                                                            Nagel's point was quite literally the opposite[1] of this, though. We can't understand what it must "be like to be a bat" because their mental model is so fundamentally different than ours. So using all the human language tokens in the world can't get us to truly understand what it's like to be a bat, or a guppy, or whatever. In fact, Nagel's point is arguably even stronger: there's no possible mental mapping between the experience of a bat and the experience of a human.

                                                                                                                            [1] https://www.sas.upenn.edu/~cavitch/pdf-library/Nagel_Bat.pdf

                                                                                                                              • Terr_

                                                                                                                                today at 6:21 AM

                                                                                                                                IMO we're a step before that: We don't even have a real fish involved, we have a character that is fictionally a fish.

                                                                                                                                In LLM-discussions, obviously-fictional characters can be useful for this, like if someone builds a "Chat with Count Dracula" app. To truly believe that a typical "AI" is some entity that "wants to be helpful" is just as mistaken as believing the same architecture creates an entity that "feels the dark thirst for the blood of the living."

                                                                                                                                Or, in this case, that it really enjoys food-pellets.

                                                                                                                                • andoando

                                                                                                                                  today at 6:36 AM

                                                                                                                                  Id highly disagree with that. Were all living in the same shared universe, and underlying every intelligence must be precisely an understanding of events happening in this space-time.

                                                                                                                                    • vixen99

                                                                                                                                      today at 1:18 PM

                                                                                                                                      What does 'precisely' mean? Everyone has the same understanding of events - a precise one?

                                                                                                                                        • andoando

                                                                                                                                          today at 4:14 PM

                                                                                                                                          No I am saying the basis of intelligence must be shared, not that we have the same exact mental model.

                                                                                                                                          I might for example say a human entered a building, a bat might on the other hand think "some big block with two sticks moved through a hole", but both are experiencing a shared physical observation, and there is some mapping between the two.

                                                                                                                                          Its like when people say, if there are aliens they would find the same mathematical constants thet we do

                                                                                                                                  • today at 4:43 AM

                                                                                                                                    • AndrewKemendo

                                                                                                                                      today at 2:37 AM

                                                                                                                                      Different argument

                                                                                                                                      I’m not going to argue other than to say that you need to view the point from a third party perspective evaluating ā€œfishā€ vs ā€œmore verbose thing,ā€ such that the composition is the determinant of the complexity of interaction (which has a unique qualia per nagel)

                                                                                                                                      Hence why it’s a ā€œunintentional nodā€ not an instantiation

                                                                                                                                        • today at 4:43 AM

                                                                                                                              • hughw

                                                                                                                                today at 12:40 PM

                                                                                                                                Tiny LLM is an oxymoron, just sayin.

                                                                                                                                  • uxcolumbo

                                                                                                                                    today at 12:58 PM

                                                                                                                                    How about: LLMs are on a spectrum and this one is on the tiny side?

                                                                                                                                    • armanified

                                                                                                                                      today at 12:47 PM

                                                                                                                                      True, but most would ignore LM if it weren't LLM.

                                                                                                                                  • gdzie-jest-sol

                                                                                                                                    today at 9:04 AM

                                                                                                                                    * How creating dataset? I download it but it is commpresed in binary format.

                                                                                                                                    * How training. In cloud or in my own dev

                                                                                                                                    * How creating a gguf

                                                                                                                                      • freetonik

                                                                                                                                        today at 9:09 AM

                                                                                                                                        You sound like Guppy. Nice touch.

                                                                                                                                        • gdzie-jest-sol

                                                                                                                                          today at 9:05 AM

                                                                                                                                          ``` uv run python -m guppylm chat

                                                                                                                                          Traceback (most recent call last):

                                                                                                                                            File "<frozen runpy>", line 198, in _run_module_as_main
                                                                                                                                            File "<frozen runpy>", line 88, in _run_code
                                                                                                                                            File "/home/user/gupik/guppylm/guppylm/__main__.py", line 48, in <module>
                                                                                                                                              main()
                                                                                                                                            File "/home/user/gupik/guppylm/guppylm/__main__.py", line 29, in main
                                                                                                                                              engine = GuppyInference("checkpoints/best_model.pt", "data/tokenizer.json")
                                                                                                                                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                                                                                                                                            File "/home/user/gupik/guppylm/guppylm/inference.py", line 17, in __init__
                                                                                                                                              self.tokenizer = Tokenizer.from_file(tokenizer_path)
                                                                                                                                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                                                                                                                                          Exception: No such file or directory (os error 2) ```

                                                                                                                                            • gdzie-jest-sol

                                                                                                                                              today at 9:49 AM

                                                                                                                                              meybe add training again (read best od fine) and train again

                                                                                                                                              ``` # after config device checkpoint_path = "checkpoints/best_model.pt"

                                                                                                                                              ckpt = torch.load(checkpoint_path, map_location=device, weights_only=False)

                                                                                                                                              model = GuppyLM(mc).to(device) if "model_state_dict" in ckpt: model.load_state_dict(ckpt["model_state_dict"]) else: model.load_state_dict(ckpt)

                                                                                                                                              start_step = ckpt.get("step", 0) print(f"Encore {start_step}") ```

                                                                                                                                      • oyebenny

                                                                                                                                        today at 5:17 AM

                                                                                                                                        Neat!

                                                                                                                                        • Elengal

                                                                                                                                          today at 9:58 AM

                                                                                                                                          Cool

                                                                                                                                          • today at 3:17 AM

                                                                                                                                            • meidad_g

                                                                                                                                              today at 4:16 PM

                                                                                                                                              [dead]

                                                                                                                                              • techpulselab

                                                                                                                                                today at 4:14 PM

                                                                                                                                                [dead]

                                                                                                                                                • textai2026

                                                                                                                                                  today at 1:52 PM

                                                                                                                                                  [dead]

                                                                                                                                                  • maxothex

                                                                                                                                                    today at 4:01 PM

                                                                                                                                                    [dead]

                                                                                                                                                    • adamsilvacons

                                                                                                                                                      today at 2:00 PM

                                                                                                                                                      [dead]

                                                                                                                                                      • solsafe_dev

                                                                                                                                                        today at 3:07 PM

                                                                                                                                                        [dead]

                                                                                                                                                        • zephyrwhimsy

                                                                                                                                                          today at 5:55 PM

                                                                                                                                                          [dead]

                                                                                                                                                          • agdexai

                                                                                                                                                            today at 1:56 PM

                                                                                                                                                            [dead]

                                                                                                                                                            • Morpheus_Matrix

                                                                                                                                                              today at 2:13 AM

                                                                                                                                                              [flagged]

                                                                                                                                                              • ethanmacavoy

                                                                                                                                                                today at 3:35 AM

                                                                                                                                                                [flagged]

                                                                                                                                                                • Alexzoofficial

                                                                                                                                                                  today at 6:14 AM

                                                                                                                                                                  [flagged]

                                                                                                                                                                  • peifeng07

                                                                                                                                                                    today at 7:36 AM

                                                                                                                                                                    [dead]

                                                                                                                                                                    • agenexus

                                                                                                                                                                      today at 4:07 AM

                                                                                                                                                                      [flagged]

                                                                                                                                                                      • zephyrwhimsy

                                                                                                                                                                        today at 7:37 AM

                                                                                                                                                                        [dead]

                                                                                                                                                                        • zephyrwhimsy

                                                                                                                                                                          today at 12:21 PM

                                                                                                                                                                          [dead]

                                                                                                                                                                          • novachen

                                                                                                                                                                            today at 11:51 AM

                                                                                                                                                                            [dead]

                                                                                                                                                                            • solsafe_dev

                                                                                                                                                                              today at 3:08 PM

                                                                                                                                                                              [dead]

                                                                                                                                                                              • techpulselab

                                                                                                                                                                                today at 8:14 AM

                                                                                                                                                                                [dead]

                                                                                                                                                                                • weiyong1024

                                                                                                                                                                                  today at 2:37 AM

                                                                                                                                                                                  [flagged]

                                                                                                                                                                                  • Morpheus_Matrix

                                                                                                                                                                                    today at 9:09 AM

                                                                                                                                                                                    [dead]

                                                                                                                                                                                    • aesopturtle

                                                                                                                                                                                      today at 4:33 AM

                                                                                                                                                                                      [flagged]

                                                                                                                                                                                      • george_belsky

                                                                                                                                                                                        today at 7:44 AM

                                                                                                                                                                                        [dead]

                                                                                                                                                                                        • aditya7303011

                                                                                                                                                                                          today at 5:10 AM

                                                                                                                                                                                          Did something similar last year https://github.com/aditya699/EduMOE

                                                                                                                                                                                          • zhichuanxun

                                                                                                                                                                                            today at 12:01 PM

                                                                                                                                                                                            [dead]

                                                                                                                                                                                            • aditya7303011

                                                                                                                                                                                              today at 5:10 AM

                                                                                                                                                                                              [dead]

                                                                                                                                                                                              • LeonTing1010

                                                                                                                                                                                                today at 3:55 AM

                                                                                                                                                                                                [flagged]

                                                                                                                                                                                              • areys

                                                                                                                                                                                                today at 12:23 PM

                                                                                                                                                                                                [flagged]

                                                                                                                                                                                                  • moonu

                                                                                                                                                                                                    today at 12:34 PM

                                                                                                                                                                                                    This comment seems ai-written

                                                                                                                                                                                                • jiusanzhou

                                                                                                                                                                                                  today at 7:01 AM

                                                                                                                                                                                                  [flagged]

                                                                                                                                                                                                    • ngruhn

                                                                                                                                                                                                      today at 7:10 AM

                                                                                                                                                                                                      comment smells AI written

                                                                                                                                                                                                      • 3m

                                                                                                                                                                                                        today at 7:21 AM

                                                                                                                                                                                                        AI account

                                                                                                                                                                                                    • dinkumthinkum

                                                                                                                                                                                                      today at 5:03 AM

                                                                                                                                                                                                      I think this is a nice project because it is end to end and serves its goal well. Good job! It's a good example how someone might do something similar for a specific purpose. There are other visualizers that explain different aspects of LLMs but this is a good applied example.

                                                                                                                                                                                                      • martmulx

                                                                                                                                                                                                        today at 3:33 AM

                                                                                                                                                                                                        How much training data did you end up needing for the fish personality to feel coherent? Curious what the minimum viable dataset looks like for something like this.

                                                                                                                                                                                                        • Propelloni

                                                                                                                                                                                                          today at 9:09 AM

                                                                                                                                                                                                          Great work! I still think that [1] does a better job of helping us understand how GPT and LLM work, but yours is funnier.

                                                                                                                                                                                                          Then, some criticism. I probably don't get it, but I think the HN headline does your project a disservice. Your project does not demystify anything (see below) and it diverges from your project's claim, too. Furthermore, I think you claim too much on your github. "This project exists to show that training your own language model is not magic." and then just posts a few command line statements to execute. Yeah, running a mail server is not magic, just apt-get install exim4. So, code. Looking at train_guppylm.ipynb and, oh, it's PyTorch again. I'm better off reading [2] if I'm looking into that (I know, it is a published book, but I maintain my point).

                                                                                                                                                                                                          So, in short, it does not help the initiated or the uninitiated. For the initiated it needs more detail for it to be useful, the uninitiated more context for it to be understood. Still a fun project, even if oversold.

                                                                                                                                                                                                          [1] https://spreadsheets-are-all-you-need.ai/ [2] https://github.com/rasbt/LLMs-from-scratch

                                                                                                                                                                                                            • jadengeller

                                                                                                                                                                                                              today at 10:46 AM

                                                                                                                                                                                                              this comment seems to be astroturfing to sell a course