Last active
March 1, 2024 14:34
-
-
Save nheingit/472a871a01d9b4cca9d6cea3fb65818a to your computer and use it in GitHub Desktop.
here is the trimming logic for the scraped mozilla docs.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
import pandas as pd | |
import os | |
import tiktoken | |
from openai import OpenAI | |
from langchain.text_splitter import RecursiveCharacterTextSplitter | |
openai = OpenAI(api_key=os.environ['OPENAI_API_KEY']) | |
DOMAIN = "developer.mozilla.org" | |
def remove_newlines(series): | |
series = series.str.replace('\n', ' ') | |
series = series.str.replace('\\n', ' ') | |
series = series.str.replace(' ', ' ') | |
series = series.str.replace(' ', ' ') | |
return series | |
# Create a list to store the text files | |
texts = [] | |
# Get all the text files in the text directory | |
for file in os.listdir("text/" + DOMAIN + "/"): | |
# Open the file and read the text | |
with open("text/" + DOMAIN + "/" + file, "r", encoding="UTF-8") as f: | |
text = f.read() | |
# we replace the last 4 characters to get rid of .txt, and replace _ with / to generate the URLs we scraped | |
filename = file[:-4].replace('_', '/') | |
""" | |
There are a lot of contributor.txt files that got included in the scrape, this weeds them out. There are also a lot of auth required urls that have been scraped to weed out as well | |
""" | |
if filename.endswith(".txt") or 'users/fxa/login' in filename: | |
continue | |
# then we replace underscores with / to get the actual links so we can cite contributions | |
texts.append((filename, text)) | |
# Create a dataframe from the list of texts | |
df = pd.DataFrame(texts, columns=['fname', 'text']) | |
# Set the text column to be the raw text with the newlines removed | |
df['text'] = df.fname + ". " + remove_newlines(df.text) | |
df.to_csv('processed/scraped.csv') | |
# Load the cl100k_base tokenizer which is designed to work with the ada-002 model | |
tokenizer = tiktoken.get_encoding("cl100k_base") | |
df = pd.read_csv('processed/scraped.csv', index_col=0) | |
df.columns = ['title', 'text'] | |
# Tokenize the text and save the number of tokens to a new column | |
df['n_tokens'] = df.text.apply(lambda x: len(tokenizer.encode(x))) | |
chunk_size = 1000 # Max number of tokens | |
text_splitter = RecursiveCharacterTextSplitter( | |
# This could be replaced with a token counting function if needed | |
length_function=len, | |
chunk_size=chunk_size, | |
chunk_overlap=0, # No overlap between chunks | |
add_start_index=False, # We don't need start index in this case | |
) | |
shortened = [] | |
for row in df.iterrows(): | |
# If the text is None, go to the next row | |
if row[1]['text'] is None: | |
continue | |
# If the number of tokens is greater than the max number of tokens, split the text into chunks | |
if row[1]['n_tokens'] > chunk_size: | |
# Split the text using LangChain's text splitter | |
chunks = text_splitter.create_documents([row[1]['text']]) | |
# Append the content of each chunk to the 'shortened' list | |
for chunk in chunks: | |
shortened.append(chunk.page_content) | |
# Otherwise, add the text to the list of shortened texts | |
else: | |
shortened.append(row[1]['text']) | |
df = pd.DataFrame(shortened, columns=['text']) | |
df['n_tokens'] = df.text.apply(lambda x: len(tokenizer.encode(x))) | |
df['embeddings'] = df.text.apply(lambda x: openai.embeddings.create( | |
input=x, model='text-embedding-ada-002').data[0].embedding) | |
df.to_csv('processed/embeddings.csv') |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment