Skip to content

Instantly share code, notes, and snippets.

Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save rowntreerob/1a6921bf525b1af9996bd65c4d34d803 to your computer and use it in GitHub Desktop.
Save rowntreerob/1a6921bf525b1af9996bd65c4d34d803 to your computer and use it in GitHub Desktop.
mml model vs algorithm debate on chatgpt training
<!DOCTYPE html><html class="light"><head> <meta charset="utf-8"/> <title>Chat GPT: Transformers: Complex Neural Learning</title> <meta name="generator" content="chatGPT Saving Bookmark"/><style>html.dark { background-color: rgb(32,33,35); color: rgb(236,236,241);}html.light { background-color: white; color: rgb(52,53,65);}body { font-size: 16px; font-family: sans-serif; line-height: 28px; margin: 0;}body > .w-full { padding: 30px;}.flex { display: flex; max-width: 100%;}.m-auto { margin: auto;}.text-base { max-width: 50rem;}.gap-1 { gap: 0.25rem;}/* model name */.justify-center { justify-content: center;}.items-center { align-items: center;}.w-4 { width: 1rem;}.h-4 { height: 1rem;}/* prompt */.dark body > .w-full:nth-of-type(2n+1) { background: rgb(52,53,65);}/* response */.dark body > .w-full:nth-of-type(2n+2) { background: rgb(68,70,84);}.light body > .w-full:nth-of-type(2n+2) { background: rgb(247,247,248);}.light body > .w-full { border-bottom: 1px solid rgba(0,0,0,.1);}a, a:visited { color: #7792cd;}pre { margin: 0 0 1em 0; display: block;}pre code.hljs { margin-bottom: 1em; border-radius: 5px;}.whitespace-pre-wrap { white-space: pre-wrap;}.flex-col { flex-direction: column;}*, :after, :before { border: 0 solid #d9d9e3; box-sizing: border-box;}table { border-collapse: collapse; border-color: inherit; text-indent: 0;}.markdown table { --tw-border-spacing-x: 0px; --tw-border-spacing-y: 0px; border-collapse: separate; border-spacing: var(--tw-border-spacing-x) var(--tw-border-spacing-y); width: 100%}.markdown th { background-color: rgba(236,236,241,.2); border-bottom-width: 1px; border-left-width: 1px; border-top-width: 1px; padding: .25rem .75rem}.markdown th:first-child { border-top-left-radius: .375rem}.markdown th:last-child { border-right-width: 1px; border-top-right-radius: .375rem}.markdown td { border-bottom-width: 1px; border-left-width: 1px; padding: .25rem .75rem}.markdown td:last-child { border-right-width: 1px}.markdown tbody tr:last-child td:first-child { border-bottom-left-radius: .375rem}.markdown tbody tr:last-child td:last-child { border-bottom-right-radius: .375rem}/* chatGPT code color theme */code.hljs,code[class*=language-],pre[class*=language-]{word-wrap:normal;background:none;color:#fff;-webkit-hyphens:none;hyphens:none;line-height:1.5;tab-size:4;text-align:left;white-space:pre;word-break:normal;word-spacing:normal}pre[class*=language-]{border-radius:.3em;overflow:auto}:not(pre)>code.hljs,:not(pre)>code[class*=language-]{border-radius:.3em;padding:.1em;white-space:normal}.hljs-comment{color:hsla(0,0%,100%,.5)}.hljs-meta{color:hsla(0,0%,100%,.6)}.hljs-built_in,.hljs-class .hljs-title{color:#e9950c}.hljs-doctag,.hljs-formula,.hljs-keyword,.hljs-literal{color:#2e95d3}.hljs-addition,.hljs-attribute,.hljs-meta-string,.hljs-regexp,.hljs-string{color:#00a67d}.hljs-attr,.hljs-number,.hljs-selector-attr,.hljs-selector-class,.hljs-selector-pseudo,.hljs-template-variable,.hljs-type,.hljs-variable{color:#df3079}.hljs-bullet,.hljs-link,.hljs-selector-id,.hljs-symbol,.hljs-title{color:#f22c3d}.token.cdata,.token.comment,.token.doctype,.token.prolog{color:#a9aec1}.token.punctuation{color:#fefefe}.token.constant,.token.deleted,.token.property,.token.symbol,.token.tag{color:#ffa07a}.token.boolean,.token.number{color:#00e0e0}.token.attr-name,.token.builtin,.token.char,.token.inserted,.token.selector,.token.string{color:#abe338}.language-css .token.string,.style .token.string,.token.entity,.token.operator,.token.url,.token.variable{color:#00e0e0}.token.atrule,.token.attr-value,.token.function{color:gold}.token.keyword{color:#00e0e0}.token.important,.token.regex{color:gold}.token.bold,.token.important{font-weight:700}.token.italic{font-style:italic}.token.entity{cursor:help}@media screen and (-ms-high-contrast:active){code[class*=language-],pre[class*=language-]{background:window;color:windowText}:not(pre)>code[class*=language-],pre[class*=language-]{background:window}.token.important{background:highlight;color:window;font-weight:400}.token.atrule,.token.attr-value,.token.function,.token.keyword,.token.operator,.token.selector{font-weight:700}.token.attr-value,.token.comment,.token.doctype,.token.function,.token.keyword,.token.operator,.token.property,.token.string{color:highlight}.token.attr-value,.token.url{font-weight:400}}/* avatars */.w-6 { width: 1.5rem;}.h-6 { height: 1.5rem;}.p-1 { padding: 0.25rem;}.w-\[30px\] { width: 30px; min-width: 30px;}.h-\[30px\] { height: 30px;}.items-end { margin: 1em 1em 0 -1em;}.items-end img { width: 100%;}/* user avatar don't have p tag with margin */body > .w-full:nth-of-type(2n+1) .items-end { margin-top: 0;}/* style of the code snippets */.rounded-md { border-radius: 0.375rem;}.mb-4 { margin-bottom: 1rem;}.p-4 { padding: 1rem;}.py-2 { padding-bottom: 0.5rem; padding-top: 0.5rem;}.px-4 { padding-left: 1rem; padding-right: 1rem;}.text-xs { font-size: .75rem; line-height: 1rem;}.bg-black { background-color: rgb(0,0,0);}.text-gray-200 { color: rgb(217,217,227);}.bg-gray-800 { background-color: rgba(52,53,65);}.rounded-t-md { border-top-left-radius: 0.375rem; border-top-right-radius: 0.375rem;}code.hljs, code[class*=language-], pre[class*=language-] { word-wrap: normal; background: none; color: #fff; -webkit-hyphens: none; hyphens: none; line-height: 1.5; tab-size: 4; text-align: left; white-space: pre; word-break: normal; word-spacing: normal;}.prose :where(code):not(:where([class~=not-prose] *)) { color: var(--tw-prose-code); font-size: .875em; font-weight: 600;}.prose :where(pre):not(:where([class~=not-prose] *)) { background-color: transparent; border-radius: 0.375rem; color: currentColor; font-size: .875em; font-weight: 400; line-height: 1.7142857; margin: 0; overflow-x: auto; padding: 0;}.prose :where(pre code):not(:where([class~=not-prose] *)) { background-color: transparent; border-radius: 0; border-width: 0; color: inherit; font-family: inherit; font-size: inherit; font-weight: inherit; line-height: inherit; padding: 0;}.\!whitespace-pre { white-space: pre!important;}.overflow-y-auto { overflow-y: auto;}.toggle { position: fixed; top: 5px; right: 5px; font-size: 16px; line-height: 1.2em;}#toggle { display: none;}#toggle + label::before { content: "☀️"; background: black; display: block; box-sizing: border-box; width: 28px; height: 28px; padding: 4px 3px; border: 1px solid white; border-radius: 50%;}#toggle:checked + label::before { content: "🌙";}</style><link rel="stylesheet" href="https://cdn.jsdelivr.net/npm/katex@0.16.4/dist/katex.min.css"/></head><body><span class="flex w-full items-center justify-center gap-1 border-b border-black/10 bg-gray-50 p-3 text-gray-500 dark:border-gray-900/50 dark:bg-gray-700 dark:text-gray-300"><svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 16 16" fill="none" class="h-4 w-4" width="16" height="16" stroke-width="2"><path d="M12.784 1.442a.8.8 0 0 0-1.569 0l-.191.953a.8.8 0 0 1-.628.628l-.953.19a.8.8 0 0 0 0 1.57l.953.19a.8.8 0 0 1 .628.629l.19.953a.8.8 0 0 0 1.57 0l.19-.953a.8.8 0 0 1 .629-.628l.953-.19a.8.8 0 0 0 0-1.57l-.953-.19a.8.8 0 0 1-.628-.629l-.19-.953h-.002ZM5.559 4.546a.8.8 0 0 0-1.519 0l-.546 1.64a.8.8 0 0 1-.507.507l-1.64.546a.8.8 0 0 0 0 1.519l1.64.547a.8.8 0 0 1 .507.505l.546 1.641a.8.8 0 0 0 1.519 0l.546-1.64a.8.8 0 0 1 .506-.507l1.641-.546a.8.8 0 0 0 0-1.519l-1.64-.546a.8.8 0 0 1-.507-.506L5.56 4.546Zm5.6 6.4a.8.8 0 0 0-1.519 0l-.147.44a.8.8 0 0 1-.505.507l-.441.146a.8.8 0 0 0 0 1.519l.44.146a.8.8 0 0 1 .507.506l.146.441a.8.8 0 0 0 1.519 0l.147-.44a.8.8 0 0 1 .506-.507l.44-.146a.8.8 0 0 0 0-1.519l-.44-.147a.8.8 0 0 1-.507-.505l-.146-.441Z" fill="currentColor"></path></svg>Model: GPT-4</span><div class="group w-full text-gray-800 dark:text-gray-100 border-b border-black/10 dark:border-gray-900/50 dark:bg-gray-800"><div class="flex p-4 gap-4 text-base md:gap-6 md:max-w-2xl lg:max-w-xl xl:max-w-3xl md:py-6 lg:px-0 m-auto"><div class="flex-shrink-0 flex flex-col relative items-end"><div class="w-[30px]"><div class="relative flex"><span style="box-sizing: border-box; display: inline-block; overflow: hidden; width: initial; height: initial; background: none; opacity: 1; border: 0px; margin: 0px; padding: 0px; position: relative; max-width: 100%;"><img alt="user avatar" decoding="async" data-nimg="intrinsic" class="rounded-sm"></span></div></div></div><div class="relative flex w-[calc(100%-50px)] flex-col gap-1 md:gap-3 lg:w-[calc(100%-115px)]"><div class="flex flex-grow flex-col gap-3"><div class="min-h-[20px] flex flex-col items-start gap-4 whitespace-pre-wrap break-words">my friends misunderstand neural networks based on transformers and make erroneous statements like there is just another software algorythm under the hood and to shed light on a decision made by AI all you have to do is examine or rerun the algorythm that is one and the same with the AI. I try to explain the model training process with transformers in neural net arrangement but it is complicated and difficult to describe the structure and the operation of a transformer based Neural net. Please help me describe why transformers are unique and why they are not at all like algorythms and should not be considered analogs to simple software algorythms producing deterministic output.</div></div></div></div></div><div class="group w-full text-gray-800 dark:text-gray-100 border-b border-black/10 dark:border-gray-900/50 bg-gray-50 dark:bg-[#444654]"><div class="flex p-4 gap-4 text-base md:gap-6 md:max-w-2xl lg:max-w-xl xl:max-w-3xl md:py-6 lg:px-0 m-auto"><div class="flex-shrink-0 flex flex-col relative items-end"><div class="w-[30px]"><div class="relative p-1 rounded-sm h-[30px] w-[30px] text-white flex items-center justify-center" style="background-color: rgb(171, 104, 255);"><svg width="41" height="41" viewBox="0 0 41 41" fill="none" xmlns="http://www.w3.org/2000/svg" stroke-width="1.5" class="h-6 w-6" role="img"><title>ChatGPT</title><text x="-9999" y="-9999">ChatGPT</text><path d="M37.5324 16.8707C37.9808 15.5241 38.1363 14.0974 37.9886 12.6859C37.8409 11.2744 37.3934 9.91076 36.676 8.68622C35.6126 6.83404 33.9882 5.3676 32.0373 4.4985C30.0864 3.62941 27.9098 3.40259 25.8215 3.85078C24.8796 2.7893 23.7219 1.94125 22.4257 1.36341C21.1295 0.785575 19.7249 0.491269 18.3058 0.500197C16.1708 0.495044 14.0893 1.16803 12.3614 2.42214C10.6335 3.67624 9.34853 5.44666 8.6917 7.47815C7.30085 7.76286 5.98686 8.3414 4.8377 9.17505C3.68854 10.0087 2.73073 11.0782 2.02839 12.312C0.956464 14.1591 0.498905 16.2988 0.721698 18.4228C0.944492 20.5467 1.83612 22.5449 3.268 24.1293C2.81966 25.4759 2.66413 26.9026 2.81182 28.3141C2.95951 29.7256 3.40701 31.0892 4.12437 32.3138C5.18791 34.1659 6.8123 35.6322 8.76321 36.5013C10.7141 37.3704 12.8907 37.5973 14.9789 37.1492C15.9208 38.2107 17.0786 39.0587 18.3747 39.6366C19.6709 40.2144 21.0755 40.5087 22.4946 40.4998C24.6307 40.5054 26.7133 39.8321 28.4418 38.5772C30.1704 37.3223 31.4556 35.5506 32.1119 33.5179C33.5027 33.2332 34.8167 32.6547 35.9659 31.821C37.115 30.9874 38.0728 29.9178 38.7752 28.684C39.8458 26.8371 40.3023 24.6979 40.0789 22.5748C39.8556 20.4517 38.9639 18.4544 37.5324 16.8707ZM22.4978 37.8849C20.7443 37.8874 19.0459 37.2733 17.6994 36.1501C17.7601 36.117 17.8666 36.0586 17.936 36.0161L25.9004 31.4156C26.1003 31.3019 26.2663 31.137 26.3813 30.9378C26.4964 30.7386 26.5563 30.5124 26.5549 30.2825V19.0542L29.9213 20.998C29.9389 21.0068 29.9541 21.0198 29.9656 21.0359C29.977 21.052 29.9842 21.0707 29.9867 21.0902V30.3889C29.9842 32.375 29.1946 34.2791 27.7909 35.6841C26.3872 37.0892 24.4838 37.8806 22.4978 37.8849ZM6.39227 31.0064C5.51397 29.4888 5.19742 27.7107 5.49804 25.9832C5.55718 26.0187 5.66048 26.0818 5.73461 26.1244L13.699 30.7248C13.8975 30.8408 14.1233 30.902 14.3532 30.902C14.583 30.902 14.8088 30.8408 15.0073 30.7248L24.731 25.1103V28.9979C24.7321 29.0177 24.7283 29.0376 24.7199 29.0556C24.7115 29.0736 24.6988 29.0893 24.6829 29.1012L16.6317 33.7497C14.9096 34.7416 12.8643 35.0097 10.9447 34.4954C9.02506 33.9811 7.38785 32.7263 6.39227 31.0064ZM4.29707 13.6194C5.17156 12.0998 6.55279 10.9364 8.19885 10.3327C8.19885 10.4013 8.19491 10.5228 8.19491 10.6071V19.808C8.19351 20.0378 8.25334 20.2638 8.36823 20.4629C8.48312 20.6619 8.64893 20.8267 8.84863 20.9404L18.5723 26.5542L15.206 28.4979C15.1894 28.5089 15.1703 28.5155 15.1505 28.5173C15.1307 28.5191 15.1107 28.516 15.0924 28.5082L7.04046 23.8557C5.32135 22.8601 4.06716 21.2235 3.55289 19.3046C3.03862 17.3858 3.30624 15.3413 4.29707 13.6194ZM31.955 20.0556L22.2312 14.4411L25.5976 12.4981C25.6142 12.4872 25.6333 12.4805 25.6531 12.4787C25.6729 12.4769 25.6928 12.4801 25.7111 12.4879L33.7631 17.1364C34.9967 17.849 36.0017 18.8982 36.6606 20.1613C37.3194 21.4244 37.6047 22.849 37.4832 24.2684C37.3617 25.6878 36.8382 27.0432 35.9743 28.1759C35.1103 29.3086 33.9415 30.1717 32.6047 30.6641C32.6047 30.5947 32.6047 30.4733 32.6047 30.3889V21.188C32.6066 20.9586 32.5474 20.7328 32.4332 20.5338C32.319 20.3348 32.154 20.1698 31.955 20.0556ZM35.3055 15.0128C35.2464 14.9765 35.1431 14.9142 35.069 14.8717L27.1045 10.2712C26.906 10.1554 26.6803 10.0943 26.4504 10.0943C26.2206 10.0943 25.9948 10.1554 25.7963 10.2712L16.0726 15.8858V11.9982C16.0715 11.9783 16.0753 11.9585 16.0837 11.9405C16.0921 11.9225 16.1048 11.9068 16.1207 11.8949L24.1719 7.25025C25.4053 6.53903 26.8158 6.19376 28.2383 6.25482C29.6608 6.31589 31.0364 6.78077 32.2044 7.59508C33.3723 8.40939 34.2842 9.53945 34.8334 10.8531C35.3826 12.1667 35.5464 13.6095 35.3055 15.0128ZM14.2424 21.9419L10.8752 19.9981C10.8576 19.9893 10.8423 19.9763 10.8309 19.9602C10.8195 19.9441 10.8122 19.9254 10.8098 19.9058V10.6071C10.8107 9.18295 11.2173 7.78848 11.9819 6.58696C12.7466 5.38544 13.8377 4.42659 15.1275 3.82264C16.4173 3.21869 17.8524 2.99464 19.2649 3.1767C20.6775 3.35876 22.0089 3.93941 23.1034 4.85067C23.0427 4.88379 22.937 4.94215 22.8668 4.98473L14.9024 9.58517C14.7025 9.69878 14.5366 9.86356 14.4215 10.0626C14.3065 10.2616 14.2466 10.4877 14.2479 10.7175L14.2424 21.9419ZM16.071 17.9991L20.4018 15.4978L24.7325 17.9975V22.9985L20.4018 25.4983L16.071 22.9985V17.9991Z" fill="currentColor"></path></svg></div></div></div><div class="relative flex w-[calc(100%-50px)] flex-col gap-1 md:gap-3 lg:w-[calc(100%-115px)]"><div class="flex flex-grow flex-col gap-3"><div class="min-h-[20px] flex flex-col items-start gap-4 whitespace-pre-wrap break-words"><div class="markdown prose w-full break-words dark:prose-invert light"><p>Certainly, I can help you explain this.</p><p>Firstly, it's important to understand that a neural network, such as the transformer models used in natural language processing, isn't just another software algorithm. An algorithm is a deterministic set of rules or steps to solve a problem. In contrast, a neural network is a model inspired by the human brain, and it's trained to learn patterns in data rather than following predefined rules.</p><p>In the case of transformers, they're designed to handle sequential data in tasks like translation, summarization, etc. Unlike other sequential models such as RNNs (Recurrent Neural Networks) and LSTMs (Long Short-Term Memory), transformers don't process data in order. Instead, they use a mechanism called "attention" to focus on different parts of the input sequence when generating the output sequence.</p><p>To understand why transformers aren't deterministic like an algorithm, consider the training process. Transformers, like all neural networks, are trained on vast amounts of data and adjust their internal parameters based on the patterns they find. These parameters, which can number in the millions or even billions, are initially set randomly. Thus, even with the same training data, two separate instances of the same model might end up with slightly different parameters.</p><p>Moreover, the learning process is probabilistic, meaning that the model calculates probabilities for its predictions, rather than deterministic outputs. For example, when generating text, a transformer model will calculate the probabilities of different words following a given input sequence, and then select a word based on those probabilities.</p><p>As for understanding the decisions made by a transformer model, it's not as simple as examining or rerunning the algorithm. Because the model's behavior is determined by its learned parameters, and these parameters are learned from data in a complex, nonlinear way, understanding why the model made a particular decision often requires advanced techniques like attention visualization or feature attribution, and even then, the interpretation can be challenging.</p><p>Moreover, the fact that these models are often trained on massive amounts of data, and the fact that their internal workings are high-dimensional and nonlinear, means that they can exhibit complex, emergent behaviors that aren't easily predictable from their structure or the data they were trained on.</p><p>In summary, transformer-based models are not just simple algorithms with deterministic outputs. They are complex, probabilistic, learning machines that are trained to find patterns in data, and their decisions can be difficult to interpret due to the complexity and high dimensionality of their internal workings.</p></div></div></div><div class="flex justify-between lg:block"><div class="text-gray-400 flex self-end lg:self-center justify-center mt-2 gap-2 md:gap-3 lg:gap-1 lg:absolute lg:top-0 lg:translate-x-full lg:right-0 lg:mt-0 lg:pl-2 visible"></div></div></div></div></div><div class="group w-full text-gray-800 dark:text-gray-100 border-b border-black/10 dark:border-gray-900/50 dark:bg-gray-800"><div class="flex p-4 gap-4 text-base md:gap-6 md:max-w-2xl lg:max-w-xl xl:max-w-3xl md:py-6 lg:px-0 m-auto"><div class="flex-shrink-0 flex flex-col relative items-end"><div class="w-[30px]"><div class="relative flex"><span style="box-sizing: border-box; display: inline-block; overflow: hidden; width: initial; height: initial; background: none; opacity: 1; border: 0px; margin: 0px; padding: 0px; position: relative; max-width: 100%;"><img alt="user avatar" decoding="async" data-nimg="intrinsic" class="rounded-sm"></span></div></div></div><div class="relative flex w-[calc(100%-50px)] flex-col gap-1 md:gap-3 lg:w-[calc(100%-115px)]"><div class="flex flex-grow flex-col gap-3"><div class="min-h-[20px] flex flex-col items-start gap-4 whitespace-pre-wrap break-words">continue and explain the importance of math, statistics, and vector relationships as input data flows through the training process and how the layers and layers math and token interrelationships makes it difficult to explain exactly why and how an AI like an autonomous car took a decision to swerve into a parked car</div></div></div></div></div><div class="group w-full text-gray-800 dark:text-gray-100 border-b border-black/10 dark:border-gray-900/50 bg-gray-50 dark:bg-[#444654]"><div class="flex p-4 gap-4 text-base md:gap-6 md:max-w-2xl lg:max-w-xl xl:max-w-3xl md:py-6 lg:px-0 m-auto"><div class="flex-shrink-0 flex flex-col relative items-end"><div class="w-[30px]"><div class="relative p-1 rounded-sm h-[30px] w-[30px] text-white flex items-center justify-center" style="background-color: rgb(171, 104, 255);"><svg width="41" height="41" viewBox="0 0 41 41" fill="none" xmlns="http://www.w3.org/2000/svg" stroke-width="1.5" class="h-6 w-6" role="img"><title>ChatGPT</title><text x="-9999" y="-9999">ChatGPT</text><path d="M37.5324 16.8707C37.9808 15.5241 38.1363 14.0974 37.9886 12.6859C37.8409 11.2744 37.3934 9.91076 36.676 8.68622C35.6126 6.83404 33.9882 5.3676 32.0373 4.4985C30.0864 3.62941 27.9098 3.40259 25.8215 3.85078C24.8796 2.7893 23.7219 1.94125 22.4257 1.36341C21.1295 0.785575 19.7249 0.491269 18.3058 0.500197C16.1708 0.495044 14.0893 1.16803 12.3614 2.42214C10.6335 3.67624 9.34853 5.44666 8.6917 7.47815C7.30085 7.76286 5.98686 8.3414 4.8377 9.17505C3.68854 10.0087 2.73073 11.0782 2.02839 12.312C0.956464 14.1591 0.498905 16.2988 0.721698 18.4228C0.944492 20.5467 1.83612 22.5449 3.268 24.1293C2.81966 25.4759 2.66413 26.9026 2.81182 28.3141C2.95951 29.7256 3.40701 31.0892 4.12437 32.3138C5.18791 34.1659 6.8123 35.6322 8.76321 36.5013C10.7141 37.3704 12.8907 37.5973 14.9789 37.1492C15.9208 38.2107 17.0786 39.0587 18.3747 39.6366C19.6709 40.2144 21.0755 40.5087 22.4946 40.4998C24.6307 40.5054 26.7133 39.8321 28.4418 38.5772C30.1704 37.3223 31.4556 35.5506 32.1119 33.5179C33.5027 33.2332 34.8167 32.6547 35.9659 31.821C37.115 30.9874 38.0728 29.9178 38.7752 28.684C39.8458 26.8371 40.3023 24.6979 40.0789 22.5748C39.8556 20.4517 38.9639 18.4544 37.5324 16.8707ZM22.4978 37.8849C20.7443 37.8874 19.0459 37.2733 17.6994 36.1501C17.7601 36.117 17.8666 36.0586 17.936 36.0161L25.9004 31.4156C26.1003 31.3019 26.2663 31.137 26.3813 30.9378C26.4964 30.7386 26.5563 30.5124 26.5549 30.2825V19.0542L29.9213 20.998C29.9389 21.0068 29.9541 21.0198 29.9656 21.0359C29.977 21.052 29.9842 21.0707 29.9867 21.0902V30.3889C29.9842 32.375 29.1946 34.2791 27.7909 35.6841C26.3872 37.0892 24.4838 37.8806 22.4978 37.8849ZM6.39227 31.0064C5.51397 29.4888 5.19742 27.7107 5.49804 25.9832C5.55718 26.0187 5.66048 26.0818 5.73461 26.1244L13.699 30.7248C13.8975 30.8408 14.1233 30.902 14.3532 30.902C14.583 30.902 14.8088 30.8408 15.0073 30.7248L24.731 25.1103V28.9979C24.7321 29.0177 24.7283 29.0376 24.7199 29.0556C24.7115 29.0736 24.6988 29.0893 24.6829 29.1012L16.6317 33.7497C14.9096 34.7416 12.8643 35.0097 10.9447 34.4954C9.02506 33.9811 7.38785 32.7263 6.39227 31.0064ZM4.29707 13.6194C5.17156 12.0998 6.55279 10.9364 8.19885 10.3327C8.19885 10.4013 8.19491 10.5228 8.19491 10.6071V19.808C8.19351 20.0378 8.25334 20.2638 8.36823 20.4629C8.48312 20.6619 8.64893 20.8267 8.84863 20.9404L18.5723 26.5542L15.206 28.4979C15.1894 28.5089 15.1703 28.5155 15.1505 28.5173C15.1307 28.5191 15.1107 28.516 15.0924 28.5082L7.04046 23.8557C5.32135 22.8601 4.06716 21.2235 3.55289 19.3046C3.03862 17.3858 3.30624 15.3413 4.29707 13.6194ZM31.955 20.0556L22.2312 14.4411L25.5976 12.4981C25.6142 12.4872 25.6333 12.4805 25.6531 12.4787C25.6729 12.4769 25.6928 12.4801 25.7111 12.4879L33.7631 17.1364C34.9967 17.849 36.0017 18.8982 36.6606 20.1613C37.3194 21.4244 37.6047 22.849 37.4832 24.2684C37.3617 25.6878 36.8382 27.0432 35.9743 28.1759C35.1103 29.3086 33.9415 30.1717 32.6047 30.6641C32.6047 30.5947 32.6047 30.4733 32.6047 30.3889V21.188C32.6066 20.9586 32.5474 20.7328 32.4332 20.5338C32.319 20.3348 32.154 20.1698 31.955 20.0556ZM35.3055 15.0128C35.2464 14.9765 35.1431 14.9142 35.069 14.8717L27.1045 10.2712C26.906 10.1554 26.6803 10.0943 26.4504 10.0943C26.2206 10.0943 25.9948 10.1554 25.7963 10.2712L16.0726 15.8858V11.9982C16.0715 11.9783 16.0753 11.9585 16.0837 11.9405C16.0921 11.9225 16.1048 11.9068 16.1207 11.8949L24.1719 7.25025C25.4053 6.53903 26.8158 6.19376 28.2383 6.25482C29.6608 6.31589 31.0364 6.78077 32.2044 7.59508C33.3723 8.40939 34.2842 9.53945 34.8334 10.8531C35.3826 12.1667 35.5464 13.6095 35.3055 15.0128ZM14.2424 21.9419L10.8752 19.9981C10.8576 19.9893 10.8423 19.9763 10.8309 19.9602C10.8195 19.9441 10.8122 19.9254 10.8098 19.9058V10.6071C10.8107 9.18295 11.2173 7.78848 11.9819 6.58696C12.7466 5.38544 13.8377 4.42659 15.1275 3.82264C16.4173 3.21869 17.8524 2.99464 19.2649 3.1767C20.6775 3.35876 22.0089 3.93941 23.1034 4.85067C23.0427 4.88379 22.937 4.94215 22.8668 4.98473L14.9024 9.58517C14.7025 9.69878 14.5366 9.86356 14.4215 10.0626C14.3065 10.2616 14.2466 10.4877 14.2479 10.7175L14.2424 21.9419ZM16.071 17.9991L20.4018 15.4978L24.7325 17.9975V22.9985L20.4018 25.4983L16.071 22.9985V17.9991Z" fill="currentColor"></path></svg></div></div></div><div class="relative flex w-[calc(100%-50px)] flex-col gap-1 md:gap-3 lg:w-[calc(100%-115px)]"><div class="flex flex-grow flex-col gap-3"><div class="min-h-[20px] flex flex-col items-start gap-4 whitespace-pre-wrap break-words"><div class="markdown prose w-full break-words dark:prose-invert light"><p>Neural networks, including transformer models, operate fundamentally on mathematical principles. Every input to the network, every transformation that occurs within the network, and every output from the network, is expressed in terms of numerical data and mathematical operations. Here's a more detailed explanation:</p><ol><li><p><strong>Input Representation:</strong> Every piece of data that's fed into a neural network, whether it's an image, a sound, or a sentence, is first transformed into a numerical representation. For transformer models dealing with text, this involves converting words or phrases (tokens) into vectors using techniques such as word embeddings. Each dimension in these vectors can represent different characteristics of the word.</p></li><li><p><strong>Mathematical Operations:</strong> As the input data flows through the network, it undergoes a series of transformations defined by mathematical functions. For instance, transformer networks heavily use linear algebra (matrix multiplications and additions), calculus (for backpropagation during training), and probability theory (in the form of softmax functions for prediction).</p></li><li><p><strong>Layers and Parameters:</strong> The structure of a neural network is composed of layers, and each layer contains numerous parameters (or weights) that are adjusted during the training process. These parameters define the relationship between the input and output of each layer.</p></li><li><p><strong>Statistics and Probability:</strong> The training process involves adjusting the network's parameters based on statistical patterns in the training data. This is typically done using a variant of stochastic gradient descent, an optimization algorithm that incrementally adjusts the parameters to minimize the difference between the model's predictions and the actual data.</p></li></ol><p>Because of this complex, multi-layered process involving high-dimensional vector spaces and non-linear transformations, it's difficult to directly relate the input and output of a transformer model. The model's decision-making process is distributed across millions or billions of parameters, and it's generally not possible to point to a single parameter or even a group of parameters as being the reason for a specific decision.</p><p>For example, if an autonomous car equipped with a transformer model were to swerve into a parked car, the decision would be the result of complex interactions between the inputs (the sensory data from the car's cameras and other sensors), the model's internal parameters (which have been trained on large amounts of data), and the model's mathematical operations. To understand why the model made the decision it did, one would need to analyze the high-dimensional vector representations and the multitude of mathematical transformations the model performed on the input data.</p><p>This complexity, combined with the fact that the training process involves randomness and that the model's behavior emerges from high-dimensional, non-linear patterns in the data, makes it difficult to provide a simple, deterministic explanation for the model's decisions. This is a major area of ongoing research in the field of AI, known as explainability or interpretability.</p></div>
<div class="fig figcenter fighighlight">
<iframe src="https://awsgcpupld-production.up.railway.app/ifrmeModel.html" height="1750px" width="850px" style="border:none;"></iframe>
</div>
</div></div><div class="flex justify-between lg:block"><div class="text-gray-400 flex self-end lg:self-center justify-center mt-2 gap-2 md:gap-3 lg:gap-1 lg:absolute lg:top-0 lg:translate-x-full lg:right-0 lg:mt-0 lg:pl-2 visible"></div></div></div></div></div><div class="toggle"><input id="toggle" type="checkbox" /><label for="toggle"></label></div><script>function decode(array) { const ua = new Uint8Array(array); return URL.createObjectURL(new Blob([ua], {type : "image/jpeg"}));}const avatar_data = { '1x': decode([255,216,255,224,0,16,74,70,73,70,0,1,1,0,0,1,0,1,0,0,255,226,1,216,73,67,67,95,80,82,79,70,73,76,69,0,1,1,0,0,1,200,0,0,0,0,4,48,0,0,109,110,116,114,82,71,66,32,88,89,90,32,7,224,0,1,0,1,0,0,0,0,0,0,97,99,115,112,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,246,214,0,1,0,0,0,0,211,45,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,9,100,101,115,99,0,0,0,240,0,0,0,36,114,88,89,90,0,0,1,20,0,0,0,20,103,88,89,90,0,0,1,40,0,0,0,20,98,88,89,90,0,0,1,60,0,0,0,20,119,116,112,116,0,0,1,80,0,0,0,20,114,84,82,67,0,0,1,100,0,0,0,40,103,84,82,67,0,0,1,100,0,0,0,40,98,84,82,67,0,0,1,100,0,0,0,40,99,112,114,116,0,0,1,140,0,0,0,60,109,108,117,99,0,0,0,0,0,0,0,1,0,0,0,12,101,110,85,83,0,0,0,8,0,0,0,28,0,115,0,82,0,71,0,66,88,89,90,32,0,0,0,0,0,0,111,162,0,0,56,245,0,0,3,144,88,89,90,32,0,0,0,0,0,0,98,153,0,0,183,133,0,0,24,218,88,89,90,32,0,0,0,0,0,0,36,160,0,0,15,132,0,0,182,207,88,89,90,32,0,0,0,0,0,0,246,214,0,1,0,0,0,0,211,45,112,97,114,97,0,0,0,0,0,4,0,0,0,2,102,102,0,0,242,167,0,0,13,89,0,0,19,208,0,0,10,91,0,0,0,0,0,0,0,0,109,108,117,99,0,0,0,0,0,0,0,1,0,0,0,12,101,110,85,83,0,0,0,32,0,0,0,28,0,71,0,111,0,111,0,103,0,108,0,101,0,32,0,73,0,110,0,99,0,46,0,32,0,50,0,48,0,49,0,54,255,219,0,67,0,2,1,1,1,1,1,2,1,1,1,2,2,2,2,2,4,3,2,2,2,2,5,4,4,3,4,6,5,6,6,6,5,6,6,6,7,9,8,6,7,9,7,6,6,8,11,8,9,10,10,10,10,10,6,8,11,12,11,10,12,9,10,10,10,255,219,0,67,1,2,2,2,2,2,2,5,3,3,5,10,7,6,7,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,255,192,0,17,8,0,48,0,48,3,1,34,0,2,17,1,3,17,1,255,196,0,27,0,0,3,0,3,1,1,0,0,0,0,0,0,0,0,0,0,7,8,9,4,5,6,1,3,255,196,0,53,16,0,1,3,3,3,2,4,5,2,4,7,0,0,0,0,0,2,1,3,4,5,6,17,0,7,33,18,19,8,49,65,97,9,20,34,50,81,21,113,22,35,161,193,38,67,82,129,146,162,194,255,196,0,24,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,5,6,4,3,7,255,196,0,32,17,0,1,4,2,3,0,3,0,0,0,0,0,0,0,0,0,1,0,3,4,5,2,33,17,18,49,19,35,52,255,218,0,12,3,1,0,2,17,3,17,0,63,0,133,255,0,163,214,73,127,149,73,144,94,253,131,79,252,235,212,162,214,253,105,78,255,0,199,71,249,86,185,48,169,254,19,112,243,254,186,166,127,190,152,159,133,167,129,75,79,196,198,236,84,93,191,40,231,26,151,68,48,104,192,93,37,57,46,151,152,2,170,241,132,81,202,162,101,21,113,162,158,150,219,13,246,42,162,13,99,179,164,6,176,209,42,127,255,0,12,92,45,142,31,161,201,111,57,199,91,120,206,177,142,13,101,140,43,35,209,157,95,15,21,31,8,79,13,204,109,211,147,173,235,41,152,79,198,135,195,209,84,209,80,151,237,35,78,165,235,76,162,229,56,207,182,53,25,60,66,237,114,109,126,225,204,183,155,101,161,142,70,70,200,182,106,170,11,156,40,46,85,112,168,168,156,123,235,36,123,22,37,56,49,241,105,182,160,145,84,6,89,229,216,32,252,186,109,73,19,184,240,145,103,95,18,128,240,135,87,175,227,91,249,49,136,87,236,227,247,214,20,158,128,199,81,255,0,77,38,52,16,14,105,23,230,239,140,201,50,21,24,18,96,67,140,43,194,68,191,186,39,151,247,255,0,109,53,63,12,95,24,80,182,135,112,229,81,67,231,94,114,90,151,101,26,107,5,33,236,174,5,5,56,76,231,205,63,174,145,83,167,132,70,240,43,213,198,72,151,205,117,208,237,69,255,0,118,109,61,239,78,220,203,77,163,55,233,51,5,197,103,200,31,79,86,201,112,188,42,42,167,175,158,136,148,209,125,174,170,222,185,204,235,100,7,28,208,87,114,187,188,219,133,191,31,41,113,92,245,119,173,107,114,156,201,197,149,74,6,5,199,70,88,30,13,20,218,69,234,203,104,133,133,85,198,112,158,171,169,201,241,117,183,44,248,16,109,42,53,149,104,200,116,150,124,201,35,87,130,207,83,110,178,235,109,138,1,175,169,43,170,36,57,194,112,73,156,170,101,131,240,125,184,22,38,249,95,179,47,214,174,193,174,29,114,24,44,234,50,72,54,127,73,124,199,164,187,141,58,88,34,36,18,30,161,225,69,48,139,198,117,159,190,20,43,22,232,241,6,213,2,152,81,231,209,32,25,209,28,97,160,66,139,45,93,140,78,186,193,39,40,66,157,177,37,84,194,161,55,194,167,168,145,176,49,165,0,66,174,183,202,60,218,178,90,60,133,38,158,178,174,23,164,53,12,41,142,19,146,13,1,164,108,122,190,165,242,21,252,47,182,184,219,138,52,202,93,106,77,14,107,125,185,16,220,237,190,25,251,75,241,167,182,175,177,251,127,97,94,213,42,80,91,34,244,138,125,65,232,253,234,131,138,243,159,73,168,249,175,174,17,50,190,186,27,110,141,165,182,183,212,49,162,191,111,70,143,45,230,92,114,4,240,15,230,3,66,88,69,235,243,201,23,90,225,115,244,142,115,198,173,195,124,133,228,46,53,190,16,122,228,133,74,101,79,228,231,71,113,179,68,233,118,59,189,124,167,229,49,198,138,219,69,177,180,177,219,81,145,92,101,214,158,172,179,221,18,238,100,155,12,170,15,60,97,85,17,21,83,211,41,248,214,159,119,152,217,43,14,229,157,71,179,118,62,149,50,35,96,218,199,145,80,184,42,4,96,184,201,38,16,145,60,254,149,84,225,83,247,224,193,183,55,10,93,182,36,122,131,244,34,105,131,97,183,26,143,9,20,198,58,18,47,210,40,74,164,65,199,11,236,188,38,184,180,192,62,167,108,172,48,147,245,226,60,92,221,129,181,117,185,127,39,92,179,42,1,68,155,14,24,129,203,25,70,211,21,30,225,25,11,82,28,5,66,109,120,33,2,95,161,50,130,68,153,69,209,239,195,182,219,111,173,221,116,82,238,184,242,89,165,83,109,202,160,1,190,25,64,80,112,145,169,8,99,254,97,43,68,97,213,148,251,243,143,69,209,120,61,184,98,208,46,216,182,205,192,226,152,200,41,20,73,230,239,2,242,162,41,176,166,75,202,116,162,160,162,168,227,43,239,195,115,107,86,235,84,186,116,173,180,219,136,141,178,108,214,198,167,54,107,65,210,174,162,2,11,13,116,47,218,233,30,84,140,209,80,16,115,141,5,57,160,220,189,37,235,36,19,7,169,241,39,62,35,39,126,147,87,185,142,166,125,130,114,171,36,36,60,190,104,223,112,186,201,83,211,8,159,246,244,198,130,85,59,161,202,253,76,221,118,43,141,53,25,132,8,2,242,96,187,107,249,79,79,45,20,247,178,229,249,173,203,184,163,84,203,165,200,239,200,12,42,245,23,204,61,32,145,122,151,242,136,138,152,247,208,99,114,46,40,177,46,195,181,232,206,61,42,171,25,224,236,210,214,56,139,142,162,138,42,8,168,240,88,246,206,169,34,254,112,164,231,31,137,242,23,255,217]), '2x': decode([255,216,255,224,0,16,74,70,73,70,0,1,1,0,0,1,0,1,0,0,255,226,1,216,73,67,67,95,80,82,79,70,73,76,69,0,1,1,0,0,1,200,0,0,0,0,4,48,0,0,109,110,116,114,82,71,66,32,88,89,90,32,7,224,0,1,0,1,0,0,0,0,0,0,97,99,115,112,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,246,214,0,1,0,0,0,0,211,45,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,9,100,101,115,99,0,0,0,240,0,0,0,36,114,88,89,90,0,0,1,20,0,0,0,20,103,88,89,90,0,0,1,40,0,0,0,20,98,88,89,90,0,0,1,60,0,0,0,20,119,116,112,116,0,0,1,80,0,0,0,20,114,84,82,67,0,0,1,100,0,0,0,40,103,84,82,67,0,0,1,100,0,0,0,40,98,84,82,67,0,0,1,100,0,0,0,40,99,112,114,116,0,0,1,140,0,0,0,60,109,108,117,99,0,0,0,0,0,0,0,1,0,0,0,12,101,110,85,83,0,0,0,8,0,0,0,28,0,115,0,82,0,71,0,66,88,89,90,32,0,0,0,0,0,0,111,162,0,0,56,245,0,0,3,144,88,89,90,32,0,0,0,0,0,0,98,153,0,0,183,133,0,0,24,218,88,89,90,32,0,0,0,0,0,0,36,160,0,0,15,132,0,0,182,207,88,89,90,32,0,0,0,0,0,0,246,214,0,1,0,0,0,0,211,45,112,97,114,97,0,0,0,0,0,4,0,0,0,2,102,102,0,0,242,167,0,0,13,89,0,0,19,208,0,0,10,91,0,0,0,0,0,0,0,0,109,108,117,99,0,0,0,0,0,0,0,1,0,0,0,12,101,110,85,83,0,0,0,32,0,0,0,28,0,71,0,111,0,111,0,103,0,108,0,101,0,32,0,73,0,110,0,99,0,46,0,32,0,50,0,48,0,49,0,54,255,219,0,67,0,2,1,1,1,1,1,2,1,1,1,2,2,2,2,2,4,3,2,2,2,2,5,4,4,3,4,6,5,6,6,6,5,6,6,6,7,9,8,6,7,9,7,6,6,8,11,8,9,10,10,10,10,10,6,8,11,12,11,10,12,9,10,10,10,255,219,0,67,1,2,2,2,2,2,2,5,3,3,5,10,7,6,7,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,255,192,0,17,8,0,96,0,96,3,1,34,0,2,17,1,3,17,1,255,196,0,28,0,0,1,5,1,1,1,0,0,0,0,0,0,0,0,0,0,6,3,4,5,7,8,9,2,1,255,196,0,58,16,0,1,3,2,5,1,6,4,4,5,3,5,0,0,0,0,1,2,3,4,5,17,0,6,7,18,33,49,8,19,34,65,81,97,20,129,145,161,50,113,177,193,21,35,162,209,225,9,51,66,68,98,210,240,241,255,196,0,27,1,0,2,2,3,1,0,0,0,0,0,0,0,0,0,0,0,5,6,1,4,0,2,7,3,255,196,0,35,17,0,2,2,2,2,2,3,1,1,1,0,0,0,0,0,0,0,1,2,3,4,17,5,18,19,33,6,49,81,65,66,35,255,218,0,12,3,1,0,2,17,3,17,0,63,0,225,121,135,2,198,204,253,240,207,186,73,86,219,156,76,181,64,175,72,240,255,0,6,124,95,167,135,30,134,68,204,151,239,21,1,96,31,81,138,170,205,160,132,171,114,34,90,142,128,144,162,191,150,61,150,210,145,112,73,183,190,36,211,149,43,104,22,92,94,70,61,177,147,51,28,231,68,72,20,226,235,202,252,8,189,175,141,60,177,139,53,241,217,254,81,10,166,82,234,130,110,70,62,54,18,202,193,85,240,73,80,210,172,204,204,85,25,148,230,197,255,0,9,73,182,26,53,146,165,189,16,153,16,195,91,250,89,87,254,216,143,52,63,79,94,153,112,214,208,197,51,162,132,242,239,219,8,63,81,54,218,24,254,172,57,123,46,157,215,19,47,238,27,255,0,56,242,40,124,91,226,111,111,251,63,206,33,73,50,84,95,250,24,46,110,229,110,238,254,248,73,249,103,139,53,253,88,126,170,62,223,250,143,232,255,0,56,240,229,36,92,30,255,0,250,63,206,54,70,175,107,211,24,247,165,67,129,143,27,85,233,231,137,3,75,72,4,247,191,211,254,112,145,140,148,155,20,116,247,196,185,30,83,246,104,199,116,223,56,1,119,171,140,40,123,167,252,225,156,141,50,148,171,25,19,216,36,95,167,150,39,164,105,228,210,57,171,145,249,61,134,79,233,242,150,222,255,0,227,111,18,47,225,219,127,223,21,60,143,65,117,94,191,128,133,67,47,37,51,76,84,201,252,60,2,63,92,106,30,199,93,130,179,118,178,132,75,166,45,254,237,229,38,239,182,45,235,229,124,82,121,19,75,102,102,140,255,0,75,162,164,46,67,14,204,66,100,128,108,74,73,28,123,121,227,185,125,142,244,218,143,148,50,5,42,145,150,32,165,182,90,136,208,117,92,13,220,27,31,215,2,179,51,29,111,67,79,199,184,104,231,88,165,163,33,79,255,0,70,220,147,6,142,234,234,181,123,203,88,254,117,193,255,0,203,25,191,180,39,250,111,215,180,210,158,252,186,115,194,92,6,46,66,203,94,95,35,237,142,231,206,203,116,134,233,46,238,98,251,135,56,205,157,166,114,4,6,41,79,193,76,29,236,205,109,126,19,209,38,222,95,92,14,163,58,217,75,83,26,114,56,44,84,190,142,4,235,38,156,179,145,51,26,41,236,2,26,113,144,180,5,32,166,222,163,223,1,79,69,218,131,97,243,198,189,237,169,164,76,69,68,181,59,79,43,145,78,89,114,49,4,130,91,62,95,166,50,108,209,221,38,192,124,134,24,49,103,220,230,156,174,53,148,100,181,37,164,67,72,98,202,252,63,124,34,227,93,60,63,124,62,125,210,146,9,23,7,166,16,117,221,214,240,253,241,122,50,96,86,54,88,41,232,47,132,138,65,55,56,118,226,82,78,213,30,70,27,184,243,109,171,105,87,56,220,206,173,154,126,102,112,84,97,225,165,180,163,110,10,205,255,0,108,66,78,212,148,58,118,20,178,1,232,148,38,216,169,42,153,154,77,76,129,50,160,235,118,38,193,165,218,255,0,158,18,166,207,98,36,160,243,109,110,183,93,199,149,126,103,20,183,161,145,209,209,26,3,70,179,199,193,103,86,42,104,116,178,166,86,149,39,158,188,245,253,49,215,142,199,90,220,205,79,37,199,134,253,72,33,7,109,223,60,247,126,130,223,92,112,189,90,167,93,122,91,50,228,187,188,181,125,169,221,235,239,141,155,216,251,95,181,34,102,65,107,35,208,221,40,83,138,27,29,113,207,0,181,250,155,120,112,11,144,131,118,111,244,115,248,197,222,22,151,233,216,29,75,237,65,66,211,125,59,145,92,133,9,185,143,70,3,105,223,127,219,140,101,6,53,167,51,107,115,234,170,69,154,233,109,238,173,110,182,193,205,134,4,180,142,133,158,149,157,103,104,174,124,204,82,234,113,51,59,45,185,182,90,239,182,202,221,181,35,201,35,210,254,120,185,224,103,236,185,164,180,247,104,20,28,149,14,91,116,246,144,212,133,194,72,46,146,157,220,175,215,207,239,129,73,245,123,29,149,14,77,182,81,21,45,6,161,212,115,164,214,245,6,26,100,55,90,132,224,101,79,11,134,158,23,217,231,238,125,49,201,29,89,201,85,77,55,212,90,198,74,171,20,41,216,83,86,1,109,123,129,108,159,9,189,135,150,59,11,173,153,157,189,67,203,117,26,142,85,133,34,4,216,240,213,240,237,186,237,198,242,8,74,188,175,99,250,227,145,185,247,39,234,98,53,14,66,115,166,92,124,174,167,41,106,98,166,232,33,167,249,252,41,54,234,47,210,255,0,174,15,241,54,119,191,71,56,249,109,123,222,144,8,164,182,160,82,180,143,158,16,114,50,80,67,136,55,0,220,140,26,74,211,183,230,19,26,159,13,77,75,73,33,77,47,155,225,173,27,77,170,149,170,171,20,182,165,50,158,253,123,74,193,190,223,151,158,24,84,59,125,28,243,160,15,38,70,249,14,58,142,131,139,98,37,231,86,169,43,118,252,30,131,4,250,183,149,213,167,121,133,121,97,215,110,226,9,239,21,182,215,233,249,219,2,136,87,121,208,99,20,52,201,132,148,66,246,97,56,1,113,106,252,94,86,199,213,180,164,128,80,111,246,196,188,138,96,109,222,232,187,123,1,206,220,70,204,107,122,149,28,43,144,122,219,20,118,159,161,198,120,246,69,123,67,117,185,41,155,7,91,226,220,115,141,17,216,35,180,43,122,93,169,173,71,174,62,130,196,130,128,200,116,95,144,79,211,174,40,22,218,122,83,127,4,210,55,21,145,212,244,194,178,50,197,86,10,217,147,18,65,14,36,220,20,11,20,31,46,127,247,166,50,252,117,101,45,126,154,85,157,44,43,84,145,212,46,210,186,153,42,179,156,169,25,143,79,171,238,196,150,144,143,135,113,179,123,244,189,250,122,12,95,122,73,166,21,74,166,149,37,121,179,52,22,166,84,226,37,115,149,20,147,41,75,59,185,87,167,255,0,113,131,251,48,246,224,201,243,50,60,93,46,214,234,83,78,76,138,162,134,106,229,22,118,196,11,18,73,241,115,108,110,254,204,178,228,212,162,49,33,204,193,21,216,10,102,200,148,151,46,160,145,208,41,62,68,126,124,223,11,86,225,186,223,233,210,56,238,98,172,186,147,148,136,157,119,203,116,157,39,200,202,110,152,151,30,135,33,144,218,228,186,187,27,175,206,220,244,244,198,111,209,157,39,161,103,220,141,156,116,71,63,180,236,182,105,18,75,148,218,140,133,120,153,82,247,20,167,125,184,60,30,125,186,98,255,0,237,145,170,121,67,63,85,232,122,55,148,42,45,206,90,101,178,39,189,29,126,14,243,155,219,142,122,31,166,1,244,194,45,38,169,168,185,223,45,211,213,221,51,77,158,203,50,163,90,254,50,161,197,253,182,253,241,56,151,60,123,187,0,62,67,68,108,82,102,50,175,118,96,204,209,235,210,32,230,76,226,243,43,105,194,139,53,226,60,124,198,30,208,116,83,41,101,201,171,248,38,229,149,164,11,184,255,0,27,189,250,155,226,243,237,25,149,228,229,45,107,173,210,93,146,211,173,88,6,123,179,248,111,215,246,192,42,25,44,13,234,23,191,158,29,113,108,242,214,167,163,155,228,84,160,6,212,52,143,34,84,161,191,11,56,83,145,80,145,41,194,182,228,190,144,8,79,149,135,63,175,150,40,61,103,236,239,11,42,56,237,91,35,41,79,71,73,37,230,58,148,143,81,247,197,227,55,50,190,210,170,57,233,43,223,34,44,126,234,147,24,139,137,10,242,81,244,252,172,109,129,186,22,100,57,166,128,136,213,144,200,152,82,83,33,180,255,0,196,249,164,250,226,196,245,252,41,52,85,19,233,146,219,34,91,173,120,84,44,70,35,170,12,71,68,114,134,16,18,226,186,47,173,190,88,38,173,102,236,143,58,49,101,80,158,142,30,64,234,47,253,176,45,50,84,56,211,147,12,202,42,221,127,17,22,251,96,53,112,114,103,64,186,216,197,108,119,145,178,149,87,53,87,217,167,210,209,185,123,194,84,125,47,231,139,95,55,233,163,66,36,117,208,152,59,219,72,67,130,221,64,243,196,151,102,61,61,248,92,174,254,111,121,5,18,38,16,27,73,28,132,243,207,235,245,197,145,53,130,233,90,84,109,181,181,32,170,221,9,183,246,197,183,23,24,139,57,210,83,183,101,75,145,52,127,40,103,76,171,152,227,230,89,36,76,98,62,234,98,146,155,221,224,15,29,127,44,41,165,58,209,172,250,106,14,158,215,115,101,66,23,55,98,35,75,83,123,218,60,36,223,218,199,235,130,74,62,76,121,250,27,110,209,165,170,51,198,105,64,81,36,110,189,185,191,183,239,139,59,54,105,30,79,204,25,109,172,151,170,210,12,105,209,90,75,148,140,228,17,224,117,195,127,1,231,155,219,165,240,43,50,159,69,254,26,249,44,141,9,100,237,100,162,80,18,154,161,170,184,186,157,146,20,244,133,88,42,221,61,108,110,79,175,219,4,61,148,181,91,49,81,181,102,85,15,51,229,227,84,26,143,61,50,73,71,133,81,192,62,16,21,99,123,221,94,151,182,5,50,39,103,8,153,155,82,169,244,9,233,80,66,27,238,212,160,139,238,183,181,241,175,244,227,178,197,27,72,179,62,94,205,114,225,41,11,17,72,135,25,195,227,11,227,194,175,75,241,207,61,48,21,69,41,12,124,133,238,202,180,202,23,180,104,51,53,126,166,225,22,32,32,115,243,197,73,169,245,163,72,203,233,128,202,9,122,160,250,88,70,211,200,7,169,251,129,243,197,203,218,58,153,42,38,174,213,19,53,176,219,221,224,11,104,42,251,108,79,158,51,206,167,102,88,177,39,187,152,97,181,222,187,78,89,109,9,38,221,126,94,120,113,193,122,199,94,132,108,213,214,90,30,213,234,140,101,92,180,186,67,204,178,224,143,27,187,72,90,111,117,43,211,211,166,43,234,124,36,211,20,226,118,221,107,85,214,163,230,112,206,185,39,48,214,31,103,50,84,91,61,210,85,203,73,230,222,247,243,250,97,248,168,162,67,183,9,229,92,226,203,123,6,15,179,167,98,56,121,22,68,122,118,119,215,136,148,247,93,96,41,9,252,91,189,122,43,202,227,235,129,218,150,131,246,122,160,119,79,70,237,117,4,158,119,15,132,71,63,115,131,14,211,185,185,188,253,68,166,181,87,167,186,185,112,226,22,100,78,123,172,130,0,29,124,192,253,241,68,86,91,136,196,69,73,254,30,207,131,200,39,21,235,169,64,57,147,157,41,122,209,171,114,13,107,37,196,151,26,159,149,171,102,125,57,81,208,202,37,133,30,169,4,95,105,233,123,250,158,152,32,204,20,151,147,21,196,52,144,74,135,242,214,158,135,20,175,102,233,75,119,39,198,134,226,19,226,93,130,210,44,69,177,116,150,115,109,29,148,154,43,233,150,149,39,249,140,189,225,227,216,243,143,86,189,3,172,185,207,236,175,244,146,77,82,177,69,171,101,218,252,155,74,166,45,111,53,100,126,33,237,244,28,251,99,102,118,125,109,26,205,161,140,211,107,8,142,205,86,154,247,141,249,45,239,32,129,198,222,69,175,140,114,244,199,242,238,166,34,63,193,136,174,205,94,201,12,21,92,4,216,155,123,245,198,128,236,121,169,113,233,122,161,63,42,213,166,110,131,83,9,76,69,19,193,82,111,127,159,35,3,114,210,208,75,137,159,92,148,94,90,7,167,153,109,58,197,76,45,198,82,43,209,84,26,121,101,62,23,211,255,0,21,167,219,175,215,7,90,131,174,52,138,78,189,35,47,80,224,170,176,184,137,238,196,212,181,252,152,170,226,233,191,61,125,125,176,11,170,121,198,38,87,155,79,175,53,37,20,186,203,8,91,49,4,84,216,184,146,155,5,147,230,57,189,173,243,196,222,69,210,154,130,178,36,122,101,38,170,166,221,171,75,50,230,76,238,183,57,222,14,164,29,222,123,143,211,11,246,122,152,205,127,253,25,152,187,66,230,19,35,84,51,54,97,171,73,220,234,148,167,64,191,226,181,248,251,226,148,155,75,129,39,36,49,86,171,36,124,93,65,254,253,77,14,169,9,60,15,153,63,108,23,118,190,171,76,165,234,21,82,147,9,221,206,137,187,18,146,45,222,90,247,253,112,25,88,153,223,76,131,77,41,184,12,12,53,241,146,83,198,20,185,56,56,100,104,134,205,142,57,9,149,196,153,72,147,200,176,81,110,215,192,49,173,70,166,207,219,32,216,35,160,245,193,70,187,234,37,75,79,234,44,56,229,63,227,91,113,162,183,28,217,254,223,75,125,127,108,86,207,55,35,87,107,233,115,34,50,166,65,4,200,14,114,16,60,189,61,241,105,253,130,79,255,217])};document.querySelectorAll('img').forEach(img => { img.src = avatar_data['2x']; img.srcset = `${avatar_data['1x']} 1x, ${avatar_data['2x']} 2x`;});toggle.addEventListener('change', () => { const className = toggle.checked ? 'dark' : 'light'; document.documentElement.className = className;});</script></body></html>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment