Last update: Tue Jan 14 23:15:49 UTC 2020 by @luckylittle
- Understand, identify, and work with containerization features
- Deploy a preconfigured application and identify crucial features such as namespaces, SELinux labels, and cgroups
DATABASE_URL=postgres://saleor:password@db/saleor | |
DEFAULT_FROM_EMAIL=noreply@domain.com | |
CELERY_BROKER_URL=redis://redis:6379/1 | |
JAEGER_AGENT_HOST=jaeger | |
SECRET_KEY=secret_key | |
API_URI=https://saleor-api.domain.com/graphql/ | |
ALLOWED_HOSTS=saleor-api.domain.com,saleor-mail.domain.com,saleor-store.domain.com,saleor-admin.domain.com,saleor-jaeger.domain.com | |
ALLOWED_CLIENT_HOSTS=saleor-api.domain.com,saleor-mail.domain.com,saleor-store.domain.com,saleor-admin.domain.com,saleor-jaeger.domain.com |
Last update: Tue Jan 14 23:15:49 UTC 2020 by @luckylittle
from types import FunctionType, CodeType | |
import dis | |
def assemble(ops): | |
cache = bytes([dis.opmap["CACHE"], 0]) | |
ret = b"" | |
for op, arg in ops: | |
opc = dis.opmap[op] | |
ret += bytes([opc, arg]) |
from transformers import ( | |
AutoConfig, | |
AutoTokenizer, | |
BitsAndBytesConfig, | |
GenerationConfig, | |
AutoModelForCausalLM, | |
LlamaTokenizerFast, | |
PreTrainedModel, | |
TextIteratorStreamer, | |
StoppingCriteria, |
aka what i did to get from nothing to done.
note: these are designed to be primarily a re-install guide for myself (writing things down helps me memorize the knowledge), as such don't take any of this on blind faith - some areas are well tested and the docs are very robust, some items, less so). YMMV
"""QA Chatbot streaming using FastAPI, LangChain Expression Language , OpenAI, and Chroma. | |
Features | |
-------- | |
- Persistent Chat Memory: | |
Stores chat history in a local file. | |
- Persistent Vector Store: | |
Stores document embeddings in a local vector store. | |
- Standalone Question Generation: | |
Rephrases follow-up questions to standalone questions in their original language. |
javascript: (function() { | |
var scripts = document.getElementsByTagName("script"), | |
regex = /(?<=(\"|\%27|\`))\/[a-zA-Z0-9_?&=\/\-\#\.]*(?=(\"|\'|\%60))/g; | |
const results = new Set; | |
for (var i = 0; i < scripts.length; i++) { | |
var t = scripts[i].src; | |
"" != t && fetch(t).then(function(t) { | |
return t.text() | |
}).then(function(t) { | |
var e = t.matchAll(regex); |
[ | |
{ | |
"name": "Sample Hoppscotch Environment - 1", | |
"variables": [ | |
{ | |
"value": "https://echo.hoppscotch.io", | |
"key": "baseURL" | |
}, | |
{ | |
"value": "var1", |
# This script requires to have some basic Python skills | |
# - Install python dependencies (thanks to TitwitMuffbiscuit on reddit) : | |
# pip install nltk beautifulsoup4 googlesearch-python trafilatura wolframalpha | |
# | |
# If you get this error "Resource punkt not found", it's because Punkt sentence tokenizer for Natural Language Toolkit is missing. | |
# Edit the file and add this before | |
# from nltk.tokenize import word_tokenize , | |
# it will download the necessary english.pickle: | |
# import nltk | |
# nltk.download('punkt') |