show dbs
By default, EBPF programs will not run on WSL2 due to required kernel modules missing. The following example error is an | |
indication of this problem: | |
modprobe: ERROR: ../libkmod/libkmod.c:586 kmod_search_moddep() could not open moddep file '/lib/modules/4.19.84-microso | |
ft-standard/modules.dep.bin' | |
modprobe: FATAL: Module kheaders not found in directory /lib/modules/4.19.84-microsoft-standard | |
chdir(/lib/modules/4.19.84-microsoft-standard/build): No such file or directory | |
To fix this you need to rebuild the WSL2 kernel with the missing kernel modules. The below instructions are for Ubuntu 18.04 WSL2. | |
1. git clone https://github.com/microsoft/WSL2-Linux-Kernel.git |
""" | |
Pickle2JSON is a simple Python Command Line program for converting Pickle file to JSON file. | |
Arguments: Only one (1) argument is expected which is the pickle file. | |
Usage: python pickle2json.py myfile.pkl | |
Output: The output is a JSON file bearing the same filename containing the JSON document of the converted Pickle file. | |
""" | |
# import libraries | |
import pickle |
CRITICAL! Almost all USDC liquidity on the REKT/USDC uniswap pool can be stolen, due to an authorization issue with burnFrom()
on the REKT token.
Uniswap v2 pools get the prices for their swaps by comparing the relative amounts of each of the two tokens that they hold. If the pool holds very little of token A, and a lot of token B, then it only takes a little of token A to buy a lot of token B.
Currently REKT and USDC are fairly priced in the pool. If there were to suddenly be very little REKT in the pool, but the same amount of USDC, then very little REKT would be able to buy a lot of USDC.
Not super comprehensive (yet), but I think having up to date documentation like this should be quite helpful for those out of the loop. Things change all the time in local AI circles, and it can be dizzying to catch up from an outsider's perspective, especially if you are new to the more technical aspects of language models in general (and not just locally hosted LLMs).
- A language model series created by Meta. Llama 1 was originally leaked in February 2023; Llama 2 then officially released later that year with openly available model weights & a permissive license. Kicked off the initial wave of open source developments that have been made when it comes to open source language modeling. The Llama series comes in four distinct sizes: 7b, 13b, 34b (only Code Llama was released for Llama 2 34b), and 70b. As of writing, the hotly anticipated Llama 3 has yet to arrive.
- Mistral AI is a French company that also distributes open weight