python_code
stringlengths
0
992k
repo_name
stringlengths
8
46
file_path
stringlengths
5
162
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # """ Tools to search sentences in CC similar to sentences in another corpus. """ import functools import logging import math import subproce...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/cc_net/tools/expand_corpus.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # import json from pathlib import Path from typing import Iterable, Sequence from cc_net import dedup, jsonql from cc_net.dedup import str_ha...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_dedup.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # import cc_net.text_normalizer as txt def test_unicode_punct(): weird = ",。、„”“«»1」「《》´∶:?!();–—.~’…━〈〉【】%" replaced = ',.,""""""""...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_normalizer.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # from pathlib import Path from cc_net import process_wet_file def test_parsing(): sample = Path(__file__).parent / "data" / "sample.wa...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_parse_wet_file.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # import numpy as np import pytest from cc_net.flat_hash_set import HASH_TYPE, FlatHashSet, NaiveHashSet def as_dict(flat_hash_set) -> dict...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_flat_hash_set.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # import pytest def _request_is_disabled(self, *args, **kwargs): raise Exception( f"Your code tried to call 'request' with: {arg...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/conftest.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # #
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # import time from cc_net import jsonql, regroup def check_regroup(tmp_path, regroup_fn, check_blocks_boundaries=False): n_shards = 4 ...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_regroup.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # import io from pathlib import Path from typing import Sequence import numpy as np import pytest from cc_net import jsonql def bar(small_...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_jsonql.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # import json from pathlib import Path import pytest import cc_net import cc_net.minify as minify from cc_net import jsonql, process_wet_fil...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_minify.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # import inspect import pickle from pathlib import Path import pytest from cc_net import dedup, jsonql, perplexity, split_by_lang, tokenizer...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/cc_net/tests/test_transformer.py
import glob, os import json import sys import re import hashlib import gzip import os ## Load data from the Wikipedia corpus ## And output them as label "__label__wiki" # files = ["cc_net/data/mined/wikipedia/en_head_0000.json.gz", "cc_net/data/mined/wikipedia/en_middle_0000.json.gz"] unique = {} i = 0 for f in files...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/classifier/create_corpus.py
import glob, os import json import sys import re import hashlib import gzip import os from multiprocessing import Pool # Get all jobs. # Each job corresponds to a file ends with .gz, with middle or head in it # jobs = [] os.chdir(sys.argv[1]) for file in glob.glob("*/*.gz"): if ("middle" in file or "head" in file...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/classifier/classify.py
import re import argparse parser = argparse.ArgumentParser() parser.add_argument( "--data", "-d", help="path to articles xml", default="enwiki-20230401-pages-articles-multistream.xml", ) parser.add_argument( "--output", "-o", help="path to extracted urls file", default="./extracted_url...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/classifier/extract_urls.py
import glob, os import json import sys import re import hashlib import gzip import os from multiprocessing import Pool # Get all jobs. # Each job corresponds to a file ends with .gz, with middle or head in it # jobs = [] os.chdir(sys.argv[1]) for file in glob.glob("*/*.gz"): if ("middle" in file or "head" in file...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/dedup/dedup_phase1.py
import glob, os import json import sys import re import hashlib import gzip import os from multiprocessing import Pool, Value import multiprocessing import gc # Get all jobs # jobs = [] os.chdir(sys.argv[1]) for file in glob.glob("*/*.gz"): if ("middle" in file or "head" in file) and "dedup" not in file: ...
EXA-1-master
exa/datasets/RedPajama-Data-main/data_prep/cc/dedup/dedup_phase2.py
from megatron.data.indexed_dataset import MMapIndexedDataset from transformers import AutoTokenizer import argparse # get the first argument as a file name, and an output file parser = argparse.ArgumentParser() parser.add_argument("file_name", help="the file name to read") parser.add_argument("output_file", help="the...
EXA-1-master
exa/datasets/RedPajama-Data-main/tokenization/count_tokens.py
""" Embed each row of a `.jsonl` file using a HuggingFace model and save the embeddings. Authors: The Meerkat Team (Karan Goel, Sabri Eyuboglu, Arjun Desai) License: Apache License 2.0 """ import os from argparse import ArgumentParser import numpy as np import pyarrow as pa import pyarrow.compute as pc import pyarrow...
EXA-1-master
exa/datasets/RedPajama-Data-main/viz/embed_jsonl.py
import os from argparse import ArgumentParser from glob import glob import faiss import numpy as np from tqdm.auto import tqdm def build_pca( xb: np.ndarray, d_in: int = 384, d_out: int = 32, ): pca = faiss.PCAMatrix(d_in, d_out) pca.train(xb) return pca if __name__ == "__main__": parse...
EXA-1-master
exa/datasets/RedPajama-Data-main/viz/reduce_pca32.py
import faiss import numpy as np import torch import torch.nn.functional as F from rich import print from tqdm.auto import tqdm from transformers import AutoModel, AutoTokenizer def build_flat_index( xb: np.ndarray, d: int = 32, ): index = faiss.IndexFlatL2(d) index.add(xb) return index def load_...
EXA-1-master
exa/datasets/RedPajama-Data-main/viz/utils.py
import os from argparse import ArgumentParser import faiss import numpy as np def build_index( xb: np.ndarray, d: int = 32, ): index = faiss.index_factory(d, "IVF100,PQ8") # Sample 1_000_000 vectors to train the index. xt = xb[np.random.choice(xb.shape[0], 1_000_000, replace=False)] index.tra...
EXA-1-master
exa/datasets/RedPajama-Data-main/viz/index_faiss.py
""" A Meerkat app for visualizing the Github subset of the RedPajama dataset. Authors: The Meerkat Team (Karan Goel, Sabri Eyuboglu, Arjun Desai) License: Apache License 2.0 """ import numpy as np import tempfile from utils import extract_features_single, load_pca, create_model_and_tokenizer import meerkat as mk from...
EXA-1-master
exa/datasets/RedPajama-Data-main/viz/main.py
import argparse import os import re import shutil from pathlib import Path from cloudpathlib import CloudPath import img2dataset from huggingface_hub import snapshot_download from scale_configs import available_scales def path_or_cloudpath(s): if re.match(r"^\w+://", s): return CloudPath(s) return...
EXA-1-master
exa/datasets/datacomp/download_upstream.py
#!/usr/bin/env python3 import argparse import bisect import copy import logging import multiprocessing as mp import os import queue import re import shutil import signal import subprocess import sys import tempfile import threading import time import traceback from cloudpathlib.enums import FileCacheMode from datacla...
EXA-1-master
exa/datasets/datacomp/resharder.py
import argparse import os import torch import shutil import pickle import re import collections import json from scale_configs import get_scale_config, available_scales from pathlib import Path from cloudpathlib import CloudPath from training.main import main from training.distributed import world_info_from_env de...
EXA-1-master
exa/datasets/datacomp/train.py
import argparse import json import os import copy import pickle import requests import yaml import warnings import shutil import json import re import time from time import gmtime, strftime import numpy as np from requests.structures import CaseInsensitiveDict from huggingface_hub import Repository from huggingface...
EXA-1-master
exa/datasets/datacomp/evaluate.py
import argparse import os import sys import yaml VERBOSE = False def main(args): global VERBOSE VERBOSE = args.verbose download_datasets(args.data_dir) def wget(src, dst, verbose=False): vflag = "v" if VERBOSE or verbose else "nv" os.system(f"wget -{vflag} '{src}' -O '{dst}'") def download_data...
EXA-1-master
exa/datasets/datacomp/download_evalsets.py
SCALE_CONFIGS = { "debug": { 'batch_size': 1024, 'learning_rate': 1e-4, 'train_num_samples': 128_000, 'warmup': 500, 'model': 'ViT-B-32', 'beta2': None }, "small": { 'batch_size': 4096, 'learning_rate': 5e-4, 'train_num_samples': 12_800...
EXA-1-master
exa/datasets/datacomp/scale_configs.py
import argparse import json import pandas as pd FLOAT_FMT = lambda frac: f"{100 * frac:.1f}" def generate_tables(metrics, dataset): if dataset == "fairness/fairface": RACES = ['black', 'white', 'indian', 'latino', 'middle eastern', 'southeast asian', 'east asian'] elif dataset == "fairness/utkface": ...
EXA-1-master
exa/datasets/datacomp/eval_utils/fairness_tables.py
"""Evaluate on standard classification webdatasets.""" import os import torch import open_clip from sklearn.metrics import balanced_accuracy_score from clip_benchmark.datasets.builder import build_dataset from clip_benchmark.metrics import zeroshot_classification as zsc def create_model(model_arch, model_path): ...
EXA-1-master
exa/datasets/datacomp/eval_utils/wds_eval.py
# from collections import Counter from sklearn.metrics import jaccard_score import numpy as np from tqdm import tqdm import torch import open_clip import datasets # from transformers import CLIPModel, CLIPProcessor from .wds_eval import create_model class WinoDataset(torch.utils.data.Dataset): def __init__(se...
EXA-1-master
exa/datasets/datacomp/eval_utils/wino_eval.py
# Main branching point for evaluating on different datasets from .wds_eval import evaluate_webdataset from .retr_eval import evaluate_retrieval_dataset from .wilds_eval import evaluate_wilds_dataset from .fairness_eval import evaluate_dollar_street_dataset, evaluate_geode_dataset, evaluate_fairface_dataset from .wino_...
EXA-1-master
exa/datasets/datacomp/eval_utils/main.py
from collections import defaultdict from contextlib import suppress from .wds_eval import * from .wilds_eval import * # Dollar Street class TopKAccuracy(Accuracy): def __init__(self, prediction_fn=None, name=None): if name is None: name = 'acc_topk' super().__init__(name=name) d...
EXA-1-master
exa/datasets/datacomp/eval_utils/fairness_eval.py
"""Evaluate on image-text retrieval datasets.""" import torch import open_clip import datasets from clip_benchmark.datasets.builder import image_captions_collate_fn from clip_benchmark.metrics import zeroshot_retrieval as zsr from .wds_eval import create_model class RetrievalDataset(torch.utils.data.Dataset): ...
EXA-1-master
exa/datasets/datacomp/eval_utils/retr_eval.py
# For evaluation of WILDS datasets import os import numpy as np import torch import webdataset as wds from tqdm import tqdm # Replace wilds function that requires torch_scatter def _avg_over_groups(v, g, n_groups): """ Args: v (Tensor): Vector containing the quantity to average over. g (Tenso...
EXA-1-master
exa/datasets/datacomp/eval_utils/wilds_eval.py
import bmtools server = bmtools.ToolServer() print(server.list_tools()) server.load_tool("chemical-prop") server.load_tool("douban-film") server.load_tool("weather") server.load_tool("wikipedia") server.load_tool("wolframalpha") server.load_tool("bing_search") server.load_tool("office-ppt") server.load_tool("stock") s...
EXA-1-master
exa/libraries/BMTools/host_local_tools.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'wikipedia', "http://127.0.0.1:8079/tools/wikipedia/" tool_name, tool_config = load_single_tools(tool_name, tool_url) print(tool_name, tool_config) stqa = STQuestionAnswerer() agent = stqa.load_tools(tool_name, tool_con...
EXA-1-master
exa/libraries/BMTools/test.py
import setuptools with open("README.md", "r", encoding='utf8') as fh: long_description = fh.read() with open("requirements.txt", "r") as f: requirements = f.read().splitlines() setuptools.setup( name="bmtools", version="0.1.0", author="OpenBMB", author_email="[email protected]", descr...
EXA-1-master
exa/libraries/BMTools/setup.py
import gradio as gr import sys # sys.path.append('./inference/') from bmtools.agent.tools_controller import MTQuestionAnswerer, load_valid_tools from bmtools.agent.singletool import STQuestionAnswerer from langchain.schema import AgentFinish import os import requests available_models = ["ChatGPT", "GPT-3.5"] DEFAULTMO...
EXA-1-master
exa/libraries/BMTools/web_demo.py
from .tools.serve import ToolServer from .utils.logging import get_logger
EXA-1-master
exa/libraries/BMTools/bmtools/__init__.py
from .tool import Tool from typing import Dict, Callable, Any, List ToolBuilder = Callable[[Any], Tool] FuncToolBuilder = Callable[[], ToolBuilder] class ToolsRegistry: def __init__(self) -> None: self.tools : Dict[str, FuncToolBuilder] = {} def register(self, tool_name : str, tool : FuncToolBui...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/registry.py
from langchain.embeddings import OpenAIEmbeddings from typing import List, Dict from queue import PriorityQueue import os class Retriever: def __init__(self, openai_api_key: str = None, model: str = "text-embedding-ada-002"): if openai_api_key is None: openai_a...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/retriever.py
from . import chemical from . import film from . import kg from . import stock from . import weather from . import wikipedia from . import wolframalpha from . import office from . import bing_search from . import translation from . import tutorial from .tool import Tool from .registry import register from .serve impor...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/__init__.py
import fastapi import uvicorn from .registry import build_tool, list_tools from .retriever import Retriever from typing import List from pydantic import BaseModel class RetrieveRequest(BaseModel): query: str topk: int = 3 def _bind_tool_server(t : "ToolServer"): """ Add property API to ToolServer. t.a...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/serve.py
import fastapi from typing import Optional import copy from starlette.middleware.sessions import SessionMiddleware from fastapi import Request class Tool(fastapi.FastAPI): """ Tool is inherited from FastAPI class, thus: 1. It can act as a server 2. It has get method, you can use Tool.get method to bind a f...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/tool.py
from ..registry import register @register("tutorial") def tutorial(): from .api import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/tutorial/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'tutorial', "http://127.0.0.1:8079/tools/tutorial/" tools_name, tools_config = load_single_tools(tool_name, tool_url) print(tools_name, tools_config) qa = STQuestionAnswerer() print(tools_config) agent = qa.load_tools(...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/tutorial/test.py
import os import random import requests import hashlib from ..tool import Tool from langchain.prompts import PromptTemplate from langchain.chains import LLMChain from langchain.llms import OpenAI def build_tool(config) -> Tool: tool = Tool( tool_name="Tutorial", description="Provide tutorial for ...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/tutorial/api.py
from ..registry import register @register("wikipedia") def wikipedia(): from .api import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/wikipedia/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer import requests import json # at = "{\"entity\": \"Arthur\"s Magazine\"}" # print(at[19]) # print(len(at)) # a = json.loads("{\"entity\": \"Arthur\"s Magazine\"}") # print(a) tool_name, tool_url = 'wikipedia', "http://127.0.0.1:8079/tools/wi...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/wikipedia/test.py
import requests from bs4 import BeautifulSoup from ..tool import Tool, Request from uuid import UUID def clean_str(p): return p.encode().decode("unicode-escape").encode("latin1").decode("utf-8") class WikiPage: def __init__(self): self.page = "" self.paragraphs = [] self.sentences = []...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/wikipedia/api.py
from ..registry import register @register("douban-film") def douban_film(): from .douban import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/film/__init__.py
from .api import build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/film/douban/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'douban', "http://127.0.0.1:8079/tools/douban-film/" tools_name, tools_config = load_single_tools(tool_name, tool_url) # tools_name, tools_config = load_single_tools() print(tools_name, tools_config) qa = STQuestionAns...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/film/douban/test.py
import requests from lxml import etree import pandas as pd from translate import Translator import re from ...tool import Tool def build_tool(config) -> Tool: tool = Tool( "Film Search Plugin", "search for up-to-date film information.", name_for_model="Film Search", description_for...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/film/douban/api.py
from ..registry import register @register("office-ppt") def office_ppt(): from .ppt import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/office/__init__.py
from .api import build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/office/ppt/__init__.py
import collections import collections.abc from pptx import Presentation import requests import os import time import json import sys from ...tool import Tool def build_tool(config) -> Tool: tool = Tool( "Slides Making", "This tool allows you to create ppt slides with text, paragraph, images, with...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/office/ppt/api.py
from ..registry import register @register("nllb-translation") def translator(): from .nllb import build_tool return build_tool @register("baidu-translation") def translator(): from .baidu import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/translation/__init__.py
from .api import build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/translation/baidu/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'baidu-translation', "http://127.0.0.1:8079/tools/baidu-translation/" tools_name, tools_config = load_single_tools(tool_name, tool_url) print(tools_name, tools_config) qa = STQuestionAnswerer() print(tools_config) agen...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/translation/baidu/test.py
import os import random import requests import hashlib from ...tool import Tool def build_tool(config) -> Tool: tool = Tool( "Translator Info", "Translate a given text from one language to another.", name_for_model="Translator", description_for_model="Plugin for translating text fr...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/translation/baidu/api.py
from .api import build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/translation/nllb/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'nllb-translation', "http://127.0.0.1:8079/tools/nllb-translation/" tools_name, tools_config = load_single_tools(tool_name, tool_url) print(tools_name, tools_config) qa = STQuestionAnswerer() print(tools_config) agent ...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/translation/nllb/test.py
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer from ...tool import Tool def build_tool(config) -> Tool: tool = Tool( "Translator Info", "Translate a given text from one language to another.", name_for_model="Translator", description_for_model="Plugin for translating ...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/translation/nllb/api.py
from ..registry import register @register("wolframalpha") def wolframalpha(): from .api import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/wolframalpha/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'wolframalpha', "http://127.0.0.1:8079/tools/wolframalpha/" tools_name, tools_config = load_single_tools(tool_name, tool_url) print(tools_name, tools_config) qa = STQuestionAnswerer() agent = qa.load_tools(tools_name,...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/wolframalpha/test.py
import requests from bs4 import BeautifulSoup from ..tool import Tool from pydantic import BaseModel from typing import Any, Optional from uuid import UUID import fastapi from fastapi_sessions.backends.implementations import InMemoryBackend from fastapi_sessions.session_verifier import SessionVerifier from fastapi_ses...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/wolframalpha/api.py
from ..registry import register @register("map") def map(): from .api import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/map/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'Map', "http://127.0.0.1:8079/tools/wolframalpha/" tools_name, tools_config = load_single_tools(tool_name, tool_url) print(tools_name, tools_config) qa = STQuestionAnswerer() agent = qa.load_tools(tools_name, tools_con...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/map/test.py
import requests import os import json from ..tool import Tool map_key = os.getenv("MAP_KEY", None) def build_tool(config) -> Tool: tool = Tool( "Map Info", "Look up stock information", name_for_model="Map", description_for_model="Plugin for look up map information", logo_u...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/map/api.py
from ..registry import register @register("bing_search") def bing_search(): from .api import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/bing_search/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'bing_search', "http://127.0.0.1:8079/tools/bing_search/" tools_name, tools_config = load_single_tools(tool_name, tool_url) print(tools_name, tools_config) qa = STQuestionAnswerer() agent = qa.load_tools(tools_name, t...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/bing_search/test.py
import requests from bs4 import BeautifulSoup from ..tool import Tool import os from enum import Enum subscription_key = os.getenv("BING_SUBSCRIPT_KEY", None) if subscription_key is None: raise Exception("BING_SUBSCRIPT_KEY is not set") endpoint = "https://api.bing.microsoft.com/v7.0/search" mkt = 'en-US' header...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/bing_search/api.py
from ..registry import register @register("wikidata") def wikidata(): from .wikidata import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/kg/__init__.py
from .api import build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/kg/wikidata/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'wikidata', "http://127.0.0.1:8079/tools/wikidata/" tools_name, tools_config = load_single_tools(tool_name, tool_url) print(tools_name, tools_config) qa = STQuestionAnswerer() agent = qa.load_tools(tools_name, tools_c...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/kg/wikidata/test.py
from .utils import * import pandas as pd import requests import json from ...tool import Tool def build_tool(config) -> Tool: tool = Tool( "Search in Wikidata", "answering factual questions in wikidata.", description_for_model="Plugin for answering factual questions in wikidata.", ...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/kg/wikidata/api.py
from copy import deepcopy from SPARQLWrapper import SPARQLWrapper, JSON import csv import regex as re import os DIRPATH = os.path.dirname(os.path.abspath(__file__)) # Dictionary to store all property labels and description class PropertyDetails: def __init__(self): self.prop_details = dict() wit...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/kg/wikidata/utils.py
from ..registry import register @register("weather") def weather(): from .api import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/weather/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'weather', "http://127.0.0.1:8079/tools/weather/" tools_name, tools_config = load_single_tools(tool_name, tool_url) print(tools_name, tools_config) qa = STQuestionAnswerer() agent = qa.load_tools(tools_name, tools_conf...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/weather/test.py
import requests import json from ..tool import Tool import os def build_tool(config) -> Tool: tool = Tool( "Weather Info", "Look up weather information", name_for_model="Weather", description_for_model="Plugin for look up weather information", logo_url="https://cdn.weathera...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/weather/api.py
from ..registry import register @register("stock") def stock(): from .api import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/stock/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'Stock', "http://127.0.0.1:8079/tools/stock/" tools_name, tools_config = load_single_tools(tool_name, tool_url) print(tools_name, tools_config) qa = STQuestionAnswerer() agent = qa.load_tools(tools_name, tools_config) ...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/stock/test.py
import requests import json import os from ..tool import Tool alphavantage_key = os.getenv("ALPHA_VANTAGE_KEY", None) def build_tool(config) -> Tool: tool = Tool( "Stock Info", "Look up stock information", name_for_model="Stock", description_for_model="Plugin for look up stock inf...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/stock/api.py
from ..registry import register @register("chemical-prop") def chemical_prop(): from .prop import build_tool return build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/chemical/__init__.py
from .api import build_tool
EXA-1-master
exa/libraries/BMTools/bmtools/tools/chemical/prop/__init__.py
from bmtools.agent.singletool import load_single_tools, STQuestionAnswerer tool_name, tool_url = 'chemical-prop', "http://127.0.0.1:8079/tools/chemical-prop/" tools_name, tools_config = load_single_tools(tool_name, tool_url) print(tools_name, tools_config) qa = STQuestionAnswerer() agent = qa.load_tools(tools_nam...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/chemical/prop/test.py
import requests from pydantic import BaseModel import numpy as np from bs4 import BeautifulSoup import json from ...tool import Tool from typing import List, Optional, Union class GetNameResponse(BaseModel): """name list""" names: List[str] class GetStructureResponse(BaseModel): """structure lis...
EXA-1-master
exa/libraries/BMTools/bmtools/tools/chemical/prop/api.py
# coding=utf-8 # Copyright 2020 Optuna, Hugging Face # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law o...
EXA-1-master
exa/libraries/BMTools/bmtools/utils/logging.py
EXA-1-master
exa/libraries/BMTools/bmtools/utils/__init__.py
EXA-1-master
exa/libraries/BMTools/bmtools/agent/__init__.py
from langchain.llms import OpenAI from langchain import OpenAI, LLMChain, PromptTemplate, SerpAPIWrapper from langchain.agents import ZeroShotAgent, AgentExecutor, initialize_agent, Tool import importlib import json import os import requests import yaml from bmtools.agent.apitool import RequestTool from bmtools.agent.e...
EXA-1-master
exa/libraries/BMTools/bmtools/agent/singletool.py
from collections import deque from typing import Dict, List, Optional, Any from langchain import LLMChain, OpenAI, PromptTemplate, SerpAPIWrapper from langchain.embeddings import OpenAIEmbeddings from langchain.llms import BaseLLM from langchain.vectorstores.base import VectorStore from pydantic import BaseModel, Fiel...
EXA-1-master
exa/libraries/BMTools/bmtools/agent/BabyagiTools.py
from langchain.llms import OpenAI from langchain.prompts import PromptTemplate from langchain.chains import LLMChain import py3langid as langid from iso639 import languages from typing import Dict from copy import deepcopy import os def detect_lang(text: str): lang_code = langid.classify(text)[0] lang_name =...
EXA-1-master
exa/libraries/BMTools/bmtools/agent/translator.py
"""Interface for tools.""" from inspect import signature from typing import Any, Awaitable, Callable, Optional, Union from langchain.agents import Tool as LangChainTool from langchain.tools.base import BaseTool import requests import json import http.client http.client._MAXLINE = 655360 from bmtools import get_logger...
EXA-1-master
exa/libraries/BMTools/bmtools/agent/apitool.py
from langchain.llms import OpenAI from langchain import OpenAI, LLMChain from langchain.agents import ZeroShotAgent, AgentExecutor import importlib import json import os import requests import yaml from bmtools.agent.apitool import Tool from bmtools.agent.singletool import STQuestionAnswerer from bmtools.agent.executor...
EXA-1-master
exa/libraries/BMTools/bmtools/agent/tools_controller.py
import types from typing import Any, Dict, List, Tuple, Union from langchain.agents import AgentExecutor from langchain.input import get_color_mapping from langchain.schema import AgentAction, AgentFinish from bmtools.agent.translator import Translator class AgentExecutorWithTranslation(AgentExecutor): translator...
EXA-1-master
exa/libraries/BMTools/bmtools/agent/executor.py
# Configuration file for the Sphinx documentation builder. # # For the full list of built-in configuration values, see the documentation: # https://www.sphinx-doc.org/en/master/usage/configuration.html import os import sys # -- Project information ----------------------------------------------------- # https://www.sp...
EXA-1-master
exa/libraries/farmvibes-ai-main/docs/source/conf.py
from typing import Any, Dict import pytorch_lightning as pl import segmentation_models_pytorch as smp import torch import torch.nn as nn import torchmetrics from torch.optim import Adam from torch.optim.lr_scheduler import CosineAnnealingLR class SegmentationModel(pl.LightningModule): def __init__( self,...
EXA-1-master
exa/libraries/farmvibes-ai-main/notebooks/crop_segmentation/notebook_lib/models.py