python_code stringlengths 0 992k | repo_name stringlengths 8 46 | file_path stringlengths 5 162 |
|---|---|---|
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
#
"""
Tools to search sentences in CC similar to sentences in another corpus.
"""
import functools
import logging
import math
import subproce... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/cc_net/tools/expand_corpus.py |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
#
import json
from pathlib import Path
from typing import Iterable, Sequence
from cc_net import dedup, jsonql
from cc_net.dedup import str_ha... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_dedup.py |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
#
import cc_net.text_normalizer as txt
def test_unicode_punct():
weird = ",。、„”“«»1」「《》´∶:?!();–—.~’…━〈〉【】%"
replaced = ',.,""""""""... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_normalizer.py |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
#
from pathlib import Path
from cc_net import process_wet_file
def test_parsing():
sample = Path(__file__).parent / "data" / "sample.wa... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_parse_wet_file.py |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
#
import numpy as np
import pytest
from cc_net.flat_hash_set import HASH_TYPE, FlatHashSet, NaiveHashSet
def as_dict(flat_hash_set) -> dict... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_flat_hash_set.py |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
#
import pytest
def _request_is_disabled(self, *args, **kwargs):
raise Exception(
f"Your code tried to call 'request' with: {arg... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/conftest.py |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
#
#
| EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
#
import time
from cc_net import jsonql, regroup
def check_regroup(tmp_path, regroup_fn, check_blocks_boundaries=False):
n_shards = 4
... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_regroup.py |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
#
import io
from pathlib import Path
from typing import Sequence
import numpy as np
import pytest
from cc_net import jsonql
def bar(small_... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_jsonql.py |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
#
import json
from pathlib import Path
import pytest
import cc_net
import cc_net.minify as minify
from cc_net import jsonql, process_wet_fil... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_minify.py |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
#
import inspect
import pickle
from pathlib import Path
import pytest
from cc_net import dedup, jsonql, perplexity, split_by_lang, tokenizer... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_transformer.py |
import glob, os
import json
import sys
import re
import hashlib
import gzip
import os
## Load data from the Wikipedia corpus
## And output them as label "__label__wiki"
#
files = ["cc_net/data/mined/wikipedia/en_head_0000.json.gz", "cc_net/data/mined/wikipedia/en_middle_0000.json.gz"]
unique = {}
i = 0
for f in files... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/classifier/create_corpus.py |
import glob, os
import json
import sys
import re
import hashlib
import gzip
import os
from multiprocessing import Pool
# Get all jobs.
# Each job corresponds to a file ends with .gz, with middle or head in it
#
jobs = []
os.chdir(sys.argv[1])
for file in glob.glob("*/*.gz"):
if ("middle" in file or "head" in file... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/classifier/classify.py |
import re
import argparse
parser = argparse.ArgumentParser()
parser.add_argument(
"--data",
"-d",
help="path to articles xml",
default="enwiki-20230401-pages-articles-multistream.xml",
)
parser.add_argument(
"--output",
"-o",
help="path to extracted urls file",
default="./extracted_url... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/classifier/extract_urls.py |
import glob, os
import json
import sys
import re
import hashlib
import gzip
import os
from multiprocessing import Pool
# Get all jobs.
# Each job corresponds to a file ends with .gz, with middle or head in it
#
jobs = []
os.chdir(sys.argv[1])
for file in glob.glob("*/*.gz"):
if ("middle" in file or "head" in file... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/dedup/dedup_phase1.py |
import glob, os
import json
import sys
import re
import hashlib
import gzip
import os
from multiprocessing import Pool, Value
import multiprocessing
import gc
# Get all jobs
#
jobs = []
os.chdir(sys.argv[1])
for file in glob.glob("*/*.gz"):
if ("middle" in file or "head" in file) and "dedup" not in file:
... | EXA-1-master | exa/datasets/RedPajama-Data-main/data_prep/cc/dedup/dedup_phase2.py |
from megatron.data.indexed_dataset import MMapIndexedDataset
from transformers import AutoTokenizer
import argparse
# get the first argument as a file name, and an output file
parser = argparse.ArgumentParser()
parser.add_argument("file_name", help="the file name to read")
parser.add_argument("output_file", help="the... | EXA-1-master | exa/datasets/RedPajama-Data-main/tokenization/count_tokens.py |
"""
Embed each row of a `.jsonl` file using a HuggingFace model and save the embeddings.
Authors: The Meerkat Team (Karan Goel, Sabri Eyuboglu, Arjun Desai)
License: Apache License 2.0
"""
import os
from argparse import ArgumentParser
import numpy as np
import pyarrow as pa
import pyarrow.compute as pc
import pyarrow... | EXA-1-master | exa/datasets/RedPajama-Data-main/viz/embed_jsonl.py |
import os
from argparse import ArgumentParser
from glob import glob
import faiss
import numpy as np
from tqdm.auto import tqdm
def build_pca(
xb: np.ndarray,
d_in: int = 384,
d_out: int = 32,
):
pca = faiss.PCAMatrix(d_in, d_out)
pca.train(xb)
return pca
if __name__ == "__main__":
parse... | EXA-1-master | exa/datasets/RedPajama-Data-main/viz/reduce_pca32.py |
import faiss
import numpy as np
import torch
import torch.nn.functional as F
from rich import print
from tqdm.auto import tqdm
from transformers import AutoModel, AutoTokenizer
def build_flat_index(
xb: np.ndarray,
d: int = 32,
):
index = faiss.IndexFlatL2(d)
index.add(xb)
return index
def load_... | EXA-1-master | exa/datasets/RedPajama-Data-main/viz/utils.py |
import os
from argparse import ArgumentParser
import faiss
import numpy as np
def build_index(
xb: np.ndarray,
d: int = 32,
):
index = faiss.index_factory(d, "IVF100,PQ8")
# Sample 1_000_000 vectors to train the index.
xt = xb[np.random.choice(xb.shape[0], 1_000_000, replace=False)]
index.tra... | EXA-1-master | exa/datasets/RedPajama-Data-main/viz/index_faiss.py |
"""
A Meerkat app for visualizing the Github subset of the RedPajama dataset.
Authors: The Meerkat Team (Karan Goel, Sabri Eyuboglu, Arjun Desai)
License: Apache License 2.0
"""
import numpy as np
import tempfile
from utils import extract_features_single, load_pca, create_model_and_tokenizer
import meerkat as mk
from... | EXA-1-master | exa/datasets/RedPajama-Data-main/viz/main.py |
import argparse
import os
import re
import shutil
from pathlib import Path
from cloudpathlib import CloudPath
import img2dataset
from huggingface_hub import snapshot_download
from scale_configs import available_scales
def path_or_cloudpath(s):
if re.match(r"^\w+://", s):
return CloudPath(s)
return... | EXA-1-master | exa/datasets/datacomp/download_upstream.py |
#!/usr/bin/env python3
import argparse
import bisect
import copy
import logging
import multiprocessing as mp
import os
import queue
import re
import shutil
import signal
import subprocess
import sys
import tempfile
import threading
import time
import traceback
from cloudpathlib.enums import FileCacheMode
from datacla... | EXA-1-master | exa/datasets/datacomp/resharder.py |
import argparse
import os
import torch
import shutil
import pickle
import re
import collections
import json
from scale_configs import get_scale_config, available_scales
from pathlib import Path
from cloudpathlib import CloudPath
from training.main import main
from training.distributed import world_info_from_env
de... | EXA-1-master | exa/datasets/datacomp/train.py |
import argparse
import json
import os
import copy
import pickle
import requests
import yaml
import warnings
import shutil
import json
import re
import time
from time import gmtime, strftime
import numpy as np
from requests.structures import CaseInsensitiveDict
from huggingface_hub import Repository
from huggingface... | EXA-1-master | exa/datasets/datacomp/evaluate.py |
import argparse
import os
import sys
import yaml
VERBOSE = False
def main(args):
global VERBOSE
VERBOSE = args.verbose
download_datasets(args.data_dir)
def wget(src, dst, verbose=False):
vflag = "v" if VERBOSE or verbose else "nv"
os.system(f"wget -{vflag} '{src}' -O '{dst}'")
def download_data... | EXA-1-master | exa/datasets/datacomp/download_evalsets.py |
SCALE_CONFIGS = {
"debug": {
'batch_size': 1024,
'learning_rate': 1e-4,
'train_num_samples': 128_000,
'warmup': 500,
'model': 'ViT-B-32',
'beta2': None
},
"small": {
'batch_size': 4096,
'learning_rate': 5e-4,
'train_num_samples': 12_800... | EXA-1-master | exa/datasets/datacomp/scale_configs.py |
import argparse
import json
import pandas as pd
FLOAT_FMT = lambda frac: f"{100 * frac:.1f}"
def generate_tables(metrics, dataset):
if dataset == "fairness/fairface":
RACES = ['black', 'white', 'indian', 'latino', 'middle eastern', 'southeast asian', 'east asian']
elif dataset == "fairness/utkface":
... | EXA-1-master | exa/datasets/datacomp/eval_utils/fairness_tables.py |
"""Evaluate on standard classification webdatasets."""
import os
import torch
import open_clip
from sklearn.metrics import balanced_accuracy_score
from clip_benchmark.datasets.builder import build_dataset
from clip_benchmark.metrics import zeroshot_classification as zsc
def create_model(model_arch, model_path):
... | EXA-1-master | exa/datasets/datacomp/eval_utils/wds_eval.py |
# from collections import Counter
from sklearn.metrics import jaccard_score
import numpy as np
from tqdm import tqdm
import torch
import open_clip
import datasets
# from transformers import CLIPModel, CLIPProcessor
from .wds_eval import create_model
class WinoDataset(torch.utils.data.Dataset):
def __init__(se... | EXA-1-master | exa/datasets/datacomp/eval_utils/wino_eval.py |
# Main branching point for evaluating on different datasets
from .wds_eval import evaluate_webdataset
from .retr_eval import evaluate_retrieval_dataset
from .wilds_eval import evaluate_wilds_dataset
from .fairness_eval import evaluate_dollar_street_dataset, evaluate_geode_dataset, evaluate_fairface_dataset
from .wino_... | EXA-1-master | exa/datasets/datacomp/eval_utils/main.py |
from collections import defaultdict
from contextlib import suppress
from .wds_eval import *
from .wilds_eval import *
# Dollar Street
class TopKAccuracy(Accuracy):
def __init__(self, prediction_fn=None, name=None):
if name is None:
name = 'acc_topk'
super().__init__(name=name)
d... | EXA-1-master | exa/datasets/datacomp/eval_utils/fairness_eval.py |
"""Evaluate on image-text retrieval datasets."""
import torch
import open_clip
import datasets
from clip_benchmark.datasets.builder import image_captions_collate_fn
from clip_benchmark.metrics import zeroshot_retrieval as zsr
from .wds_eval import create_model
class RetrievalDataset(torch.utils.data.Dataset):
... | EXA-1-master | exa/datasets/datacomp/eval_utils/retr_eval.py |
# For evaluation of WILDS datasets
import os
import numpy as np
import torch
import webdataset as wds
from tqdm import tqdm
# Replace wilds function that requires torch_scatter
def _avg_over_groups(v, g, n_groups):
"""
Args:
v (Tensor): Vector containing the quantity to average over.
g (Tenso... | EXA-1-master | exa/datasets/datacomp/eval_utils/wilds_eval.py |
import bmtools
server = bmtools.ToolServer()
print(server.list_tools())
server.load_tool("chemical-prop")
server.load_tool("douban-film")
server.load_tool("weather")
server.load_tool("wikipedia")
server.load_tool("wolframalpha")
server.load_tool("bing_search")
server.load_tool("office-ppt")
server.load_tool("stock")
s... | EXA-1-master | exa/libraries/BMTools/host_local_tools.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'wikipedia', "http://127.0.0.1:8079/tools/wikipedia/"
tool_name, tool_config = load_single_tools(tool_name, tool_url)
print(tool_name, tool_config)
stqa = STQuestionAnswerer()
agent = stqa.load_tools(tool_name, tool_con... | EXA-1-master | exa/libraries/BMTools/test.py |
import setuptools
with open("README.md", "r", encoding='utf8') as fh:
long_description = fh.read()
with open("requirements.txt", "r") as f:
requirements = f.read().splitlines()
setuptools.setup(
name="bmtools",
version="0.1.0",
author="OpenBMB",
author_email="[email protected]",
descr... | EXA-1-master | exa/libraries/BMTools/setup.py |
import gradio as gr
import sys
# sys.path.append('./inference/')
from bmtools.agent.tools_controller import MTQuestionAnswerer, load_valid_tools
from bmtools.agent.singletool import STQuestionAnswerer
from langchain.schema import AgentFinish
import os
import requests
available_models = ["ChatGPT", "GPT-3.5"]
DEFAULTMO... | EXA-1-master | exa/libraries/BMTools/web_demo.py |
from .tools.serve import ToolServer
from .utils.logging import get_logger
| EXA-1-master | exa/libraries/BMTools/bmtools/__init__.py |
from .tool import Tool
from typing import Dict, Callable, Any, List
ToolBuilder = Callable[[Any], Tool]
FuncToolBuilder = Callable[[], ToolBuilder]
class ToolsRegistry:
def __init__(self) -> None:
self.tools : Dict[str, FuncToolBuilder] = {}
def register(self, tool_name : str, tool : FuncToolBui... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/registry.py |
from langchain.embeddings import OpenAIEmbeddings
from typing import List, Dict
from queue import PriorityQueue
import os
class Retriever:
def __init__(self,
openai_api_key: str = None,
model: str = "text-embedding-ada-002"):
if openai_api_key is None:
openai_a... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/retriever.py |
from . import chemical
from . import film
from . import kg
from . import stock
from . import weather
from . import wikipedia
from . import wolframalpha
from . import office
from . import bing_search
from . import translation
from . import tutorial
from .tool import Tool
from .registry import register
from .serve impor... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/__init__.py |
import fastapi
import uvicorn
from .registry import build_tool, list_tools
from .retriever import Retriever
from typing import List
from pydantic import BaseModel
class RetrieveRequest(BaseModel):
query: str
topk: int = 3
def _bind_tool_server(t : "ToolServer"):
""" Add property API to ToolServer.
t.a... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/serve.py |
import fastapi
from typing import Optional
import copy
from starlette.middleware.sessions import SessionMiddleware
from fastapi import Request
class Tool(fastapi.FastAPI):
""" Tool is inherited from FastAPI class, thus:
1. It can act as a server
2. It has get method, you can use Tool.get method to bind a f... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/tool.py |
from ..registry import register
@register("tutorial")
def tutorial():
from .api import build_tool
return build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/tutorial/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'tutorial', "http://127.0.0.1:8079/tools/tutorial/"
tools_name, tools_config = load_single_tools(tool_name, tool_url)
print(tools_name, tools_config)
qa = STQuestionAnswerer()
print(tools_config)
agent = qa.load_tools(... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/tutorial/test.py |
import os
import random
import requests
import hashlib
from ..tool import Tool
from langchain.prompts import PromptTemplate
from langchain.chains import LLMChain
from langchain.llms import OpenAI
def build_tool(config) -> Tool:
tool = Tool(
tool_name="Tutorial",
description="Provide tutorial for ... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/tutorial/api.py |
from ..registry import register
@register("wikipedia")
def wikipedia():
from .api import build_tool
return build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/wikipedia/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
import requests
import json
# at = "{\"entity\": \"Arthur\"s Magazine\"}"
# print(at[19])
# print(len(at))
# a = json.loads("{\"entity\": \"Arthur\"s Magazine\"}")
# print(a)
tool_name, tool_url = 'wikipedia', "http://127.0.0.1:8079/tools/wi... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/wikipedia/test.py |
import requests
from bs4 import BeautifulSoup
from ..tool import Tool, Request
from uuid import UUID
def clean_str(p):
return p.encode().decode("unicode-escape").encode("latin1").decode("utf-8")
class WikiPage:
def __init__(self):
self.page = ""
self.paragraphs = []
self.sentences = []... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/wikipedia/api.py |
from ..registry import register
@register("douban-film")
def douban_film():
from .douban import build_tool
return build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/film/__init__.py |
from .api import build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/film/douban/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'douban', "http://127.0.0.1:8079/tools/douban-film/"
tools_name, tools_config = load_single_tools(tool_name, tool_url)
# tools_name, tools_config = load_single_tools()
print(tools_name, tools_config)
qa = STQuestionAns... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/film/douban/test.py |
import requests
from lxml import etree
import pandas as pd
from translate import Translator
import re
from ...tool import Tool
def build_tool(config) -> Tool:
tool = Tool(
"Film Search Plugin",
"search for up-to-date film information.",
name_for_model="Film Search",
description_for... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/film/douban/api.py |
from ..registry import register
@register("office-ppt")
def office_ppt():
from .ppt import build_tool
return build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/office/__init__.py |
from .api import build_tool | EXA-1-master | exa/libraries/BMTools/bmtools/tools/office/ppt/__init__.py |
import collections
import collections.abc
from pptx import Presentation
import requests
import os
import time
import json
import sys
from ...tool import Tool
def build_tool(config) -> Tool:
tool = Tool(
"Slides Making",
"This tool allows you to create ppt slides with text, paragraph, images, with... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/office/ppt/api.py |
from ..registry import register
@register("nllb-translation")
def translator():
from .nllb import build_tool
return build_tool
@register("baidu-translation")
def translator():
from .baidu import build_tool
return build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/translation/__init__.py |
from .api import build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/translation/baidu/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'baidu-translation', "http://127.0.0.1:8079/tools/baidu-translation/"
tools_name, tools_config = load_single_tools(tool_name, tool_url)
print(tools_name, tools_config)
qa = STQuestionAnswerer()
print(tools_config)
agen... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/translation/baidu/test.py |
import os
import random
import requests
import hashlib
from ...tool import Tool
def build_tool(config) -> Tool:
tool = Tool(
"Translator Info",
"Translate a given text from one language to another.",
name_for_model="Translator",
description_for_model="Plugin for translating text fr... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/translation/baidu/api.py |
from .api import build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/translation/nllb/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'nllb-translation', "http://127.0.0.1:8079/tools/nllb-translation/"
tools_name, tools_config = load_single_tools(tool_name, tool_url)
print(tools_name, tools_config)
qa = STQuestionAnswerer()
print(tools_config)
agent ... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/translation/nllb/test.py |
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
from ...tool import Tool
def build_tool(config) -> Tool:
tool = Tool(
"Translator Info",
"Translate a given text from one language to another.",
name_for_model="Translator",
description_for_model="Plugin for translating ... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/translation/nllb/api.py |
from ..registry import register
@register("wolframalpha")
def wolframalpha():
from .api import build_tool
return build_tool | EXA-1-master | exa/libraries/BMTools/bmtools/tools/wolframalpha/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'wolframalpha', "http://127.0.0.1:8079/tools/wolframalpha/"
tools_name, tools_config = load_single_tools(tool_name, tool_url)
print(tools_name, tools_config)
qa = STQuestionAnswerer()
agent = qa.load_tools(tools_name,... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/wolframalpha/test.py |
import requests
from bs4 import BeautifulSoup
from ..tool import Tool
from pydantic import BaseModel
from typing import Any, Optional
from uuid import UUID
import fastapi
from fastapi_sessions.backends.implementations import InMemoryBackend
from fastapi_sessions.session_verifier import SessionVerifier
from fastapi_ses... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/wolframalpha/api.py |
from ..registry import register
@register("map")
def map():
from .api import build_tool
return build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/map/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'Map', "http://127.0.0.1:8079/tools/wolframalpha/"
tools_name, tools_config = load_single_tools(tool_name, tool_url)
print(tools_name, tools_config)
qa = STQuestionAnswerer()
agent = qa.load_tools(tools_name, tools_con... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/map/test.py |
import requests
import os
import json
from ..tool import Tool
map_key = os.getenv("MAP_KEY", None)
def build_tool(config) -> Tool:
tool = Tool(
"Map Info",
"Look up stock information",
name_for_model="Map",
description_for_model="Plugin for look up map information",
logo_u... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/map/api.py |
from ..registry import register
@register("bing_search")
def bing_search():
from .api import build_tool
return build_tool | EXA-1-master | exa/libraries/BMTools/bmtools/tools/bing_search/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'bing_search', "http://127.0.0.1:8079/tools/bing_search/"
tools_name, tools_config = load_single_tools(tool_name, tool_url)
print(tools_name, tools_config)
qa = STQuestionAnswerer()
agent = qa.load_tools(tools_name, t... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/bing_search/test.py |
import requests
from bs4 import BeautifulSoup
from ..tool import Tool
import os
from enum import Enum
subscription_key = os.getenv("BING_SUBSCRIPT_KEY", None)
if subscription_key is None:
raise Exception("BING_SUBSCRIPT_KEY is not set")
endpoint = "https://api.bing.microsoft.com/v7.0/search"
mkt = 'en-US'
header... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/bing_search/api.py |
from ..registry import register
@register("wikidata")
def wikidata():
from .wikidata import build_tool
return build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/kg/__init__.py |
from .api import build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/kg/wikidata/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'wikidata', "http://127.0.0.1:8079/tools/wikidata/"
tools_name, tools_config = load_single_tools(tool_name, tool_url)
print(tools_name, tools_config)
qa = STQuestionAnswerer()
agent = qa.load_tools(tools_name, tools_c... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/kg/wikidata/test.py |
from .utils import *
import pandas as pd
import requests
import json
from ...tool import Tool
def build_tool(config) -> Tool:
tool = Tool(
"Search in Wikidata",
"answering factual questions in wikidata.",
description_for_model="Plugin for answering factual questions in wikidata.",
... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/kg/wikidata/api.py |
from copy import deepcopy
from SPARQLWrapper import SPARQLWrapper, JSON
import csv
import regex as re
import os
DIRPATH = os.path.dirname(os.path.abspath(__file__))
# Dictionary to store all property labels and description
class PropertyDetails:
def __init__(self):
self.prop_details = dict()
wit... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/kg/wikidata/utils.py |
from ..registry import register
@register("weather")
def weather():
from .api import build_tool
return build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/weather/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'weather', "http://127.0.0.1:8079/tools/weather/"
tools_name, tools_config = load_single_tools(tool_name, tool_url)
print(tools_name, tools_config)
qa = STQuestionAnswerer()
agent = qa.load_tools(tools_name, tools_conf... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/weather/test.py |
import requests
import json
from ..tool import Tool
import os
def build_tool(config) -> Tool:
tool = Tool(
"Weather Info",
"Look up weather information",
name_for_model="Weather",
description_for_model="Plugin for look up weather information",
logo_url="https://cdn.weathera... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/weather/api.py |
from ..registry import register
@register("stock")
def stock():
from .api import build_tool
return build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/stock/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'Stock', "http://127.0.0.1:8079/tools/stock/"
tools_name, tools_config = load_single_tools(tool_name, tool_url)
print(tools_name, tools_config)
qa = STQuestionAnswerer()
agent = qa.load_tools(tools_name, tools_config)
... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/stock/test.py |
import requests
import json
import os
from ..tool import Tool
alphavantage_key = os.getenv("ALPHA_VANTAGE_KEY", None)
def build_tool(config) -> Tool:
tool = Tool(
"Stock Info",
"Look up stock information",
name_for_model="Stock",
description_for_model="Plugin for look up stock inf... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/stock/api.py |
from ..registry import register
@register("chemical-prop")
def chemical_prop():
from .prop import build_tool
return build_tool
| EXA-1-master | exa/libraries/BMTools/bmtools/tools/chemical/__init__.py |
from .api import build_tool | EXA-1-master | exa/libraries/BMTools/bmtools/tools/chemical/prop/__init__.py |
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer
tool_name, tool_url = 'chemical-prop', "http://127.0.0.1:8079/tools/chemical-prop/"
tools_name, tools_config = load_single_tools(tool_name, tool_url)
print(tools_name, tools_config)
qa = STQuestionAnswerer()
agent = qa.load_tools(tools_nam... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/chemical/prop/test.py |
import requests
from pydantic import BaseModel
import numpy as np
from bs4 import BeautifulSoup
import json
from ...tool import Tool
from typing import List, Optional, Union
class GetNameResponse(BaseModel):
"""name list"""
names: List[str]
class GetStructureResponse(BaseModel):
"""structure lis... | EXA-1-master | exa/libraries/BMTools/bmtools/tools/chemical/prop/api.py |
# coding=utf-8
# Copyright 2020 Optuna, Hugging Face
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law o... | EXA-1-master | exa/libraries/BMTools/bmtools/utils/logging.py |
EXA-1-master | exa/libraries/BMTools/bmtools/utils/__init__.py | |
EXA-1-master | exa/libraries/BMTools/bmtools/agent/__init__.py | |
from langchain.llms import OpenAI
from langchain import OpenAI, LLMChain, PromptTemplate, SerpAPIWrapper
from langchain.agents import ZeroShotAgent, AgentExecutor, initialize_agent, Tool
import importlib
import json
import os
import requests
import yaml
from bmtools.agent.apitool import RequestTool
from bmtools.agent.e... | EXA-1-master | exa/libraries/BMTools/bmtools/agent/singletool.py |
from collections import deque
from typing import Dict, List, Optional, Any
from langchain import LLMChain, OpenAI, PromptTemplate, SerpAPIWrapper
from langchain.embeddings import OpenAIEmbeddings
from langchain.llms import BaseLLM
from langchain.vectorstores.base import VectorStore
from pydantic import BaseModel, Fiel... | EXA-1-master | exa/libraries/BMTools/bmtools/agent/BabyagiTools.py |
from langchain.llms import OpenAI
from langchain.prompts import PromptTemplate
from langchain.chains import LLMChain
import py3langid as langid
from iso639 import languages
from typing import Dict
from copy import deepcopy
import os
def detect_lang(text: str):
lang_code = langid.classify(text)[0]
lang_name =... | EXA-1-master | exa/libraries/BMTools/bmtools/agent/translator.py |
"""Interface for tools."""
from inspect import signature
from typing import Any, Awaitable, Callable, Optional, Union
from langchain.agents import Tool as LangChainTool
from langchain.tools.base import BaseTool
import requests
import json
import http.client
http.client._MAXLINE = 655360
from bmtools import get_logger... | EXA-1-master | exa/libraries/BMTools/bmtools/agent/apitool.py |
from langchain.llms import OpenAI
from langchain import OpenAI, LLMChain
from langchain.agents import ZeroShotAgent, AgentExecutor
import importlib
import json
import os
import requests
import yaml
from bmtools.agent.apitool import Tool
from bmtools.agent.singletool import STQuestionAnswerer
from bmtools.agent.executor... | EXA-1-master | exa/libraries/BMTools/bmtools/agent/tools_controller.py |
import types
from typing import Any, Dict, List, Tuple, Union
from langchain.agents import AgentExecutor
from langchain.input import get_color_mapping
from langchain.schema import AgentAction, AgentFinish
from bmtools.agent.translator import Translator
class AgentExecutorWithTranslation(AgentExecutor):
translator... | EXA-1-master | exa/libraries/BMTools/bmtools/agent/executor.py |
# Configuration file for the Sphinx documentation builder.
#
# For the full list of built-in configuration values, see the documentation:
# https://www.sphinx-doc.org/en/master/usage/configuration.html
import os
import sys
# -- Project information -----------------------------------------------------
# https://www.sp... | EXA-1-master | exa/libraries/farmvibes-ai-main/docs/source/conf.py |
from typing import Any, Dict
import pytorch_lightning as pl
import segmentation_models_pytorch as smp
import torch
import torch.nn as nn
import torchmetrics
from torch.optim import Adam
from torch.optim.lr_scheduler import CosineAnnealingLR
class SegmentationModel(pl.LightningModule):
def __init__(
self,... | EXA-1-master | exa/libraries/farmvibes-ai-main/notebooks/crop_segmentation/notebook_lib/models.py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.