mirror of
https://github.com/Aider-AI/aider.git
synced 2025-05-20 20:35:00 +00:00
wip
This commit is contained in:
parent
fceaa0504c
commit
af48cc3e4c
7 changed files with 180 additions and 47 deletions
|
@ -75,6 +75,7 @@ class Coder:
|
||||||
edit_format=None,
|
edit_format=None,
|
||||||
io=None,
|
io=None,
|
||||||
from_coder=None,
|
from_coder=None,
|
||||||
|
summarize_from_coder=True,
|
||||||
**kwargs,
|
**kwargs,
|
||||||
):
|
):
|
||||||
from . import (
|
from . import (
|
||||||
|
@ -108,7 +109,7 @@ class Coder:
|
||||||
# confused the new LLM. It may try and imitate it, disobeying
|
# confused the new LLM. It may try and imitate it, disobeying
|
||||||
# the system prompt.
|
# the system prompt.
|
||||||
done_messages = from_coder.done_messages
|
done_messages = from_coder.done_messages
|
||||||
if edit_format != from_coder.edit_format and done_messages:
|
if edit_format != from_coder.edit_format and done_messages and summarize_from_coder:
|
||||||
done_messages = from_coder.summarizer.summarize_all(done_messages)
|
done_messages = from_coder.summarizer.summarize_all(done_messages)
|
||||||
|
|
||||||
# Bring along context from the old Coder
|
# Bring along context from the old Coder
|
||||||
|
@ -550,18 +551,16 @@ class Coder:
|
||||||
files_reply = "Ok, any changes I propose will be to those files."
|
files_reply = "Ok, any changes I propose will be to those files."
|
||||||
elif repo_content:
|
elif repo_content:
|
||||||
files_content = self.gpt_prompts.files_no_full_files_with_repo_map
|
files_content = self.gpt_prompts.files_no_full_files_with_repo_map
|
||||||
files_reply = (
|
files_reply = self.gpt_prompts.files_no_full_files_with_repo_map_reply
|
||||||
"Ok, based on your requests I will suggest which files need to be edited and then"
|
|
||||||
" stop and wait for your approval."
|
|
||||||
)
|
|
||||||
else:
|
else:
|
||||||
files_content = self.gpt_prompts.files_no_full_files
|
files_content = self.gpt_prompts.files_no_full_files
|
||||||
files_reply = "Ok."
|
files_reply = "Ok."
|
||||||
|
|
||||||
files_messages += [
|
if files_content:
|
||||||
dict(role="user", content=files_content),
|
files_messages += [
|
||||||
dict(role="assistant", content=files_reply),
|
dict(role="user", content=files_content),
|
||||||
]
|
dict(role="assistant", content=files_reply),
|
||||||
|
]
|
||||||
|
|
||||||
images_message = self.get_images_message()
|
images_message = self.get_images_message()
|
||||||
if images_message is not None:
|
if images_message is not None:
|
||||||
|
@ -734,7 +733,8 @@ class Coder:
|
||||||
|
|
||||||
example_messages = []
|
example_messages = []
|
||||||
if self.main_model.examples_as_sys_msg:
|
if self.main_model.examples_as_sys_msg:
|
||||||
main_sys += "\n# Example conversations:\n\n"
|
if self.gpt_prompts.example_messages:
|
||||||
|
main_sys += "\n# Example conversations:\n\n"
|
||||||
for msg in self.gpt_prompts.example_messages:
|
for msg in self.gpt_prompts.example_messages:
|
||||||
role = msg["role"]
|
role = msg["role"]
|
||||||
content = self.fmt_system_prompt(msg["content"])
|
content = self.fmt_system_prompt(msg["content"])
|
||||||
|
|
|
@ -28,6 +28,11 @@ Only include the files that are most likely to actually need to be edited.
|
||||||
Don't include files that might contain relevant context, just files that will need to be changed.
|
Don't include files that might contain relevant context, just files that will need to be changed.
|
||||||
""" # noqa: E501
|
""" # noqa: E501
|
||||||
|
|
||||||
|
files_no_full_files_with_repo_map_reply = (
|
||||||
|
"Ok, based on your requests I will suggest which files need to be edited and then"
|
||||||
|
" stop and wait for your approval."
|
||||||
|
)
|
||||||
|
|
||||||
repo_content_prefix = """Here are summaries of some files present in my git repository.
|
repo_content_prefix = """Here are summaries of some files present in my git repository.
|
||||||
Do not propose changes to these files, treat them as *read-only*.
|
Do not propose changes to these files, treat them as *read-only*.
|
||||||
If you need to edit any of these files, ask me to *add them to the chat* first.
|
If you need to edit any of these files, ask me to *add them to the chat* first.
|
||||||
|
|
|
@ -29,3 +29,15 @@ Unless the question indicates otherwise, assume the user wants to use aider as a
|
||||||
|
|
||||||
example_messages = []
|
example_messages = []
|
||||||
system_reminder = ""
|
system_reminder = ""
|
||||||
|
|
||||||
|
files_content_prefix = """These are some files we have been discussing that we may want to edit after you answer my questions:
|
||||||
|
"""
|
||||||
|
|
||||||
|
files_no_full_files = "I am not sharing any files with you."
|
||||||
|
|
||||||
|
files_no_full_files_with_repo_map = ""
|
||||||
|
files_no_full_files_with_repo_map_reply = ""
|
||||||
|
|
||||||
|
repo_content_prefix = """Here are summaries of some files present in my git repository.
|
||||||
|
We may look at these in more detail after you answer my questions.
|
||||||
|
"""
|
||||||
|
|
|
@ -656,6 +656,7 @@ class Commands:
|
||||||
io=self.io,
|
io=self.io,
|
||||||
from_coder=self.coder,
|
from_coder=self.coder,
|
||||||
edit_format="help",
|
edit_format="help",
|
||||||
|
summarize_from_coder=False,
|
||||||
)
|
)
|
||||||
user_msg = self.help.ask(args)
|
user_msg = self.help.ask(args)
|
||||||
user_msg += """
|
user_msg += """
|
||||||
|
|
|
@ -5,8 +5,8 @@ import sys
|
||||||
import warnings
|
import warnings
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
|
||||||
from tqdm import tqdm
|
|
||||||
import importlib_resources
|
import importlib_resources
|
||||||
|
from tqdm import tqdm
|
||||||
|
|
||||||
from aider.dump import dump # noqa: F401
|
from aider.dump import dump # noqa: F401
|
||||||
|
|
||||||
|
@ -14,16 +14,20 @@ warnings.simplefilter("ignore", category=FutureWarning)
|
||||||
|
|
||||||
|
|
||||||
def get_package_files():
|
def get_package_files():
|
||||||
website_files = importlib_resources.files('website')
|
for path in importlib_resources.files("website").iterdir():
|
||||||
for path in importlib_resources.files('website').iterdir():
|
dump(path)
|
||||||
if path.is_file() and path.name.endswith('.md'):
|
if path.is_file() and path.name.endswith(".md"):
|
||||||
if not any(part.startswith(('OLD', 'tmp')) or part in ('examples', '_posts') for part in path.parts):
|
if not any(
|
||||||
dump(path)
|
part.startswith(("OLD", "tmp")) or part in ("examples", "_posts")
|
||||||
|
for part in path.parts
|
||||||
|
):
|
||||||
yield str(path)
|
yield str(path)
|
||||||
elif path.is_dir():
|
elif path.is_dir():
|
||||||
for subpath in path.rglob('*.md'):
|
for subpath in path.rglob("*.md"):
|
||||||
if not any(part.startswith(('OLD', 'tmp')) or part in ('examples', '_posts') for part in subpath.parts):
|
if not any(
|
||||||
dump(subpath)
|
part.startswith(("OLD", "tmp")) or part in ("examples", "_posts")
|
||||||
|
for part in subpath.parts
|
||||||
|
):
|
||||||
yield str(subpath)
|
yield str(subpath)
|
||||||
|
|
||||||
|
|
||||||
|
@ -35,9 +39,6 @@ def fname_to_url(filepath):
|
||||||
docid = ""
|
docid = ""
|
||||||
if filepath.startswith("website/_includes/"):
|
if filepath.startswith("website/_includes/"):
|
||||||
pass
|
pass
|
||||||
elif "HISTORY.html" in filepath:
|
|
||||||
# too much stale info
|
|
||||||
pass
|
|
||||||
elif filepath.startswith(website):
|
elif filepath.startswith(website):
|
||||||
docid = filepath[len(website) :]
|
docid = filepath[len(website) :]
|
||||||
|
|
||||||
|
@ -73,8 +74,9 @@ def get_index():
|
||||||
nodes = []
|
nodes = []
|
||||||
for fname in tqdm(list(get_package_files())):
|
for fname in tqdm(list(get_package_files())):
|
||||||
fname = Path(fname)
|
fname = Path(fname)
|
||||||
|
dump(fname)
|
||||||
doc = Document(
|
doc = Document(
|
||||||
text=importlib_resources.files('website').joinpath(fname).read_text(),
|
text=importlib_resources.files("website").joinpath(fname).read_text(),
|
||||||
metadata=dict(
|
metadata=dict(
|
||||||
filename=fname.name,
|
filename=fname.name,
|
||||||
extension=fname.suffix,
|
extension=fname.suffix,
|
||||||
|
|
|
@ -29,6 +29,7 @@ watchdog
|
||||||
flake8
|
flake8
|
||||||
llama-index-core
|
llama-index-core
|
||||||
llama-index-embeddings-huggingface
|
llama-index-embeddings-huggingface
|
||||||
|
importlib_resources
|
||||||
|
|
||||||
# v3.3 no longer works on python 3.9
|
# v3.3 no longer works on python 3.9
|
||||||
networkx<3.3
|
networkx<3.3
|
||||||
|
|
160
requirements.txt
160
requirements.txt
|
@ -5,7 +5,10 @@
|
||||||
# pip-compile requirements.in
|
# pip-compile requirements.in
|
||||||
#
|
#
|
||||||
aiohttp==3.9.5
|
aiohttp==3.9.5
|
||||||
# via litellm
|
# via
|
||||||
|
# huggingface-hub
|
||||||
|
# litellm
|
||||||
|
# llama-index-core
|
||||||
aiosignal==1.3.1
|
aiosignal==1.3.1
|
||||||
# via aiohttp
|
# via aiohttp
|
||||||
altair==5.3.0
|
altair==5.3.0
|
||||||
|
@ -31,7 +34,7 @@ cachetools==5.3.3
|
||||||
# via
|
# via
|
||||||
# google-auth
|
# google-auth
|
||||||
# streamlit
|
# streamlit
|
||||||
certifi==2024.6.2
|
certifi==2024.7.4
|
||||||
# via
|
# via
|
||||||
# httpcore
|
# httpcore
|
||||||
# httpx
|
# httpx
|
||||||
|
@ -45,17 +48,27 @@ charset-normalizer==3.3.2
|
||||||
click==8.1.7
|
click==8.1.7
|
||||||
# via
|
# via
|
||||||
# litellm
|
# litellm
|
||||||
|
# nltk
|
||||||
# streamlit
|
# streamlit
|
||||||
configargparse==1.7
|
configargparse==1.7
|
||||||
# via -r requirements.in
|
# via -r requirements.in
|
||||||
|
dataclasses-json==0.6.7
|
||||||
|
# via llama-index-core
|
||||||
|
deprecated==1.2.14
|
||||||
|
# via llama-index-core
|
||||||
diff-match-patch==20230430
|
diff-match-patch==20230430
|
||||||
# via -r requirements.in
|
# via -r requirements.in
|
||||||
|
dirtyjson==1.0.8
|
||||||
|
# via llama-index-core
|
||||||
diskcache==5.6.3
|
diskcache==5.6.3
|
||||||
# via -r requirements.in
|
# via -r requirements.in
|
||||||
distro==1.9.0
|
distro==1.9.0
|
||||||
# via openai
|
# via openai
|
||||||
filelock==3.15.4
|
filelock==3.15.4
|
||||||
# via huggingface-hub
|
# via
|
||||||
|
# huggingface-hub
|
||||||
|
# torch
|
||||||
|
# transformers
|
||||||
flake8==7.1.0
|
flake8==7.1.0
|
||||||
# via -r requirements.in
|
# via -r requirements.in
|
||||||
frozenlist==1.4.1
|
frozenlist==1.4.1
|
||||||
|
@ -63,7 +76,10 @@ frozenlist==1.4.1
|
||||||
# aiohttp
|
# aiohttp
|
||||||
# aiosignal
|
# aiosignal
|
||||||
fsspec==2024.6.1
|
fsspec==2024.6.1
|
||||||
# via huggingface-hub
|
# via
|
||||||
|
# huggingface-hub
|
||||||
|
# llama-index-core
|
||||||
|
# torch
|
||||||
gitdb==4.0.11
|
gitdb==4.0.11
|
||||||
# via gitpython
|
# via gitpython
|
||||||
gitpython==3.1.43
|
gitpython==3.1.43
|
||||||
|
@ -77,9 +93,9 @@ google-api-core[grpc]==2.19.1
|
||||||
# google-ai-generativelanguage
|
# google-ai-generativelanguage
|
||||||
# google-api-python-client
|
# google-api-python-client
|
||||||
# google-generativeai
|
# google-generativeai
|
||||||
google-api-python-client==2.135.0
|
google-api-python-client==2.136.0
|
||||||
# via google-generativeai
|
# via google-generativeai
|
||||||
google-auth==2.30.0
|
google-auth==2.31.0
|
||||||
# via
|
# via
|
||||||
# google-ai-generativelanguage
|
# google-ai-generativelanguage
|
||||||
# google-api-core
|
# google-api-core
|
||||||
|
@ -95,7 +111,9 @@ googleapis-common-protos==1.63.2
|
||||||
# google-api-core
|
# google-api-core
|
||||||
# grpcio-status
|
# grpcio-status
|
||||||
greenlet==3.0.3
|
greenlet==3.0.3
|
||||||
# via playwright
|
# via
|
||||||
|
# playwright
|
||||||
|
# sqlalchemy
|
||||||
grep-ast==0.3.2
|
grep-ast==0.3.2
|
||||||
# via -r requirements.in
|
# via -r requirements.in
|
||||||
grpcio==1.64.1
|
grpcio==1.64.1
|
||||||
|
@ -113,17 +131,22 @@ httplib2==0.22.0
|
||||||
# google-api-python-client
|
# google-api-python-client
|
||||||
# google-auth-httplib2
|
# google-auth-httplib2
|
||||||
httpx==0.27.0
|
httpx==0.27.0
|
||||||
# via openai
|
# via
|
||||||
huggingface-hub==0.23.4
|
# llama-cloud
|
||||||
# via tokenizers
|
# llama-index-core
|
||||||
|
# openai
|
||||||
|
huggingface-hub[inference]==0.23.4
|
||||||
|
# via
|
||||||
|
# llama-index-embeddings-huggingface
|
||||||
|
# sentence-transformers
|
||||||
|
# tokenizers
|
||||||
|
# transformers
|
||||||
idna==3.7
|
idna==3.7
|
||||||
# via
|
# via
|
||||||
# anyio
|
# anyio
|
||||||
# httpx
|
# httpx
|
||||||
# requests
|
# requests
|
||||||
# yarl
|
# yarl
|
||||||
ijson==3.3.0
|
|
||||||
# via litellm
|
|
||||||
importlib-metadata==7.2.1
|
importlib-metadata==7.2.1
|
||||||
# via
|
# via
|
||||||
# -r requirements.in
|
# -r requirements.in
|
||||||
|
@ -133,60 +156,99 @@ jinja2==3.1.4
|
||||||
# altair
|
# altair
|
||||||
# litellm
|
# litellm
|
||||||
# pydeck
|
# pydeck
|
||||||
|
# torch
|
||||||
|
joblib==1.4.2
|
||||||
|
# via
|
||||||
|
# nltk
|
||||||
|
# scikit-learn
|
||||||
jsonschema==4.22.0
|
jsonschema==4.22.0
|
||||||
# via
|
# via
|
||||||
# -r requirements.in
|
# -r requirements.in
|
||||||
# altair
|
# altair
|
||||||
|
# litellm
|
||||||
jsonschema-specifications==2023.12.1
|
jsonschema-specifications==2023.12.1
|
||||||
# via jsonschema
|
# via jsonschema
|
||||||
litellm==1.41.0
|
litellm==1.41.6
|
||||||
|
# via -r requirements.in
|
||||||
|
llama-cloud==0.0.6
|
||||||
|
# via llama-index-core
|
||||||
|
llama-index-core==0.10.52.post2
|
||||||
|
# via
|
||||||
|
# -r requirements.in
|
||||||
|
# llama-index-embeddings-huggingface
|
||||||
|
llama-index-embeddings-huggingface==0.2.2
|
||||||
# via -r requirements.in
|
# via -r requirements.in
|
||||||
markdown-it-py==3.0.0
|
markdown-it-py==3.0.0
|
||||||
# via rich
|
# via rich
|
||||||
markupsafe==2.1.5
|
markupsafe==2.1.5
|
||||||
# via jinja2
|
# via jinja2
|
||||||
|
marshmallow==3.21.3
|
||||||
|
# via dataclasses-json
|
||||||
mccabe==0.7.0
|
mccabe==0.7.0
|
||||||
# via flake8
|
# via flake8
|
||||||
mdurl==0.1.2
|
mdurl==0.1.2
|
||||||
# via markdown-it-py
|
# via markdown-it-py
|
||||||
|
minijinja==2.0.1
|
||||||
|
# via huggingface-hub
|
||||||
|
mpmath==1.3.0
|
||||||
|
# via sympy
|
||||||
multidict==6.0.5
|
multidict==6.0.5
|
||||||
# via
|
# via
|
||||||
# aiohttp
|
# aiohttp
|
||||||
# yarl
|
# yarl
|
||||||
|
mypy-extensions==1.0.0
|
||||||
|
# via typing-inspect
|
||||||
|
nest-asyncio==1.6.0
|
||||||
|
# via llama-index-core
|
||||||
networkx==3.2.1
|
networkx==3.2.1
|
||||||
# via -r requirements.in
|
# via
|
||||||
numpy==2.0.0
|
# -r requirements.in
|
||||||
|
# llama-index-core
|
||||||
|
# torch
|
||||||
|
nltk==3.8.1
|
||||||
|
# via llama-index-core
|
||||||
|
numpy==1.26.4
|
||||||
# via
|
# via
|
||||||
# -r requirements.in
|
# -r requirements.in
|
||||||
# altair
|
# altair
|
||||||
|
# llama-index-core
|
||||||
# pandas
|
# pandas
|
||||||
# pyarrow
|
# pyarrow
|
||||||
# pydeck
|
# pydeck
|
||||||
|
# scikit-learn
|
||||||
# scipy
|
# scipy
|
||||||
|
# sentence-transformers
|
||||||
# streamlit
|
# streamlit
|
||||||
openai==1.35.7
|
# transformers
|
||||||
|
openai==1.35.10
|
||||||
# via
|
# via
|
||||||
# -r requirements.in
|
# -r requirements.in
|
||||||
# litellm
|
# litellm
|
||||||
|
# llama-index-core
|
||||||
packaging==24.1
|
packaging==24.1
|
||||||
# via
|
# via
|
||||||
# -r requirements.in
|
# -r requirements.in
|
||||||
# altair
|
# altair
|
||||||
# huggingface-hub
|
# huggingface-hub
|
||||||
|
# marshmallow
|
||||||
# streamlit
|
# streamlit
|
||||||
|
# transformers
|
||||||
pandas==2.2.2
|
pandas==2.2.2
|
||||||
# via
|
# via
|
||||||
# altair
|
# altair
|
||||||
|
# llama-index-core
|
||||||
# streamlit
|
# streamlit
|
||||||
pathspec==0.12.1
|
pathspec==0.12.1
|
||||||
# via
|
# via
|
||||||
# -r requirements.in
|
# -r requirements.in
|
||||||
# grep-ast
|
# grep-ast
|
||||||
pillow==10.3.0
|
pillow==10.4.0
|
||||||
# via
|
# via
|
||||||
# -r requirements.in
|
# -r requirements.in
|
||||||
|
# llama-index-core
|
||||||
|
# sentence-transformers
|
||||||
# streamlit
|
# streamlit
|
||||||
playwright==1.44.0
|
playwright==1.45.0
|
||||||
# via -r requirements.in
|
# via -r requirements.in
|
||||||
prompt-toolkit==3.0.47
|
prompt-toolkit==3.0.47
|
||||||
# via -r requirements.in
|
# via -r requirements.in
|
||||||
|
@ -215,12 +277,13 @@ pycodestyle==2.12.0
|
||||||
# via flake8
|
# via flake8
|
||||||
pycparser==2.22
|
pycparser==2.22
|
||||||
# via cffi
|
# via cffi
|
||||||
pydantic==2.7.4
|
pydantic==2.8.2
|
||||||
# via
|
# via
|
||||||
# google-generativeai
|
# google-generativeai
|
||||||
# litellm
|
# litellm
|
||||||
|
# llama-cloud
|
||||||
# openai
|
# openai
|
||||||
pydantic-core==2.18.4
|
pydantic-core==2.20.1
|
||||||
# via pydantic
|
# via pydantic
|
||||||
pydeck==0.9.1
|
pydeck==0.9.1
|
||||||
# via streamlit
|
# via streamlit
|
||||||
|
@ -244,19 +307,26 @@ pyyaml==6.0.1
|
||||||
# via
|
# via
|
||||||
# -r requirements.in
|
# -r requirements.in
|
||||||
# huggingface-hub
|
# huggingface-hub
|
||||||
|
# llama-index-core
|
||||||
|
# transformers
|
||||||
referencing==0.35.1
|
referencing==0.35.1
|
||||||
# via
|
# via
|
||||||
# jsonschema
|
# jsonschema
|
||||||
# jsonschema-specifications
|
# jsonschema-specifications
|
||||||
regex==2024.5.15
|
regex==2024.5.15
|
||||||
# via tiktoken
|
# via
|
||||||
|
# nltk
|
||||||
|
# tiktoken
|
||||||
|
# transformers
|
||||||
requests==2.32.3
|
requests==2.32.3
|
||||||
# via
|
# via
|
||||||
# google-api-core
|
# google-api-core
|
||||||
# huggingface-hub
|
# huggingface-hub
|
||||||
# litellm
|
# litellm
|
||||||
|
# llama-index-core
|
||||||
# streamlit
|
# streamlit
|
||||||
# tiktoken
|
# tiktoken
|
||||||
|
# transformers
|
||||||
rich==13.7.1
|
rich==13.7.1
|
||||||
# via
|
# via
|
||||||
# -r requirements.in
|
# -r requirements.in
|
||||||
|
@ -267,8 +337,17 @@ rpds-py==0.18.1
|
||||||
# referencing
|
# referencing
|
||||||
rsa==4.9
|
rsa==4.9
|
||||||
# via google-auth
|
# via google-auth
|
||||||
|
safetensors==0.4.3
|
||||||
|
# via transformers
|
||||||
|
scikit-learn==1.5.1
|
||||||
|
# via sentence-transformers
|
||||||
scipy==1.13.1
|
scipy==1.13.1
|
||||||
# via -r requirements.in
|
# via
|
||||||
|
# -r requirements.in
|
||||||
|
# scikit-learn
|
||||||
|
# sentence-transformers
|
||||||
|
sentence-transformers==3.0.1
|
||||||
|
# via llama-index-embeddings-huggingface
|
||||||
six==1.16.0
|
six==1.16.0
|
||||||
# via python-dateutil
|
# via python-dateutil
|
||||||
smmap==5.0.1
|
smmap==5.0.1
|
||||||
|
@ -284,27 +363,48 @@ soundfile==0.12.1
|
||||||
# via -r requirements.in
|
# via -r requirements.in
|
||||||
soupsieve==2.5
|
soupsieve==2.5
|
||||||
# via beautifulsoup4
|
# via beautifulsoup4
|
||||||
|
sqlalchemy[asyncio]==2.0.31
|
||||||
|
# via
|
||||||
|
# llama-index-core
|
||||||
|
# sqlalchemy
|
||||||
streamlit==1.36.0
|
streamlit==1.36.0
|
||||||
# via -r requirements.in
|
# via -r requirements.in
|
||||||
|
sympy==1.12.1
|
||||||
|
# via torch
|
||||||
tenacity==8.4.2
|
tenacity==8.4.2
|
||||||
# via streamlit
|
# via
|
||||||
|
# llama-index-core
|
||||||
|
# streamlit
|
||||||
|
threadpoolctl==3.5.0
|
||||||
|
# via scikit-learn
|
||||||
tiktoken==0.7.0
|
tiktoken==0.7.0
|
||||||
# via
|
# via
|
||||||
# -r requirements.in
|
# -r requirements.in
|
||||||
# litellm
|
# litellm
|
||||||
|
# llama-index-core
|
||||||
tokenizers==0.19.1
|
tokenizers==0.19.1
|
||||||
# via litellm
|
# via
|
||||||
|
# litellm
|
||||||
|
# transformers
|
||||||
toml==0.10.2
|
toml==0.10.2
|
||||||
# via streamlit
|
# via streamlit
|
||||||
toolz==0.12.1
|
toolz==0.12.1
|
||||||
# via altair
|
# via altair
|
||||||
|
torch==2.2.2
|
||||||
|
# via sentence-transformers
|
||||||
tornado==6.4.1
|
tornado==6.4.1
|
||||||
# via streamlit
|
# via streamlit
|
||||||
tqdm==4.66.4
|
tqdm==4.66.4
|
||||||
# via
|
# via
|
||||||
# google-generativeai
|
# google-generativeai
|
||||||
# huggingface-hub
|
# huggingface-hub
|
||||||
|
# llama-index-core
|
||||||
|
# nltk
|
||||||
# openai
|
# openai
|
||||||
|
# sentence-transformers
|
||||||
|
# transformers
|
||||||
|
transformers==4.42.3
|
||||||
|
# via sentence-transformers
|
||||||
tree-sitter==0.21.3
|
tree-sitter==0.21.3
|
||||||
# via
|
# via
|
||||||
# -r requirements.in
|
# -r requirements.in
|
||||||
|
@ -315,11 +415,19 @@ typing-extensions==4.12.2
|
||||||
# via
|
# via
|
||||||
# google-generativeai
|
# google-generativeai
|
||||||
# huggingface-hub
|
# huggingface-hub
|
||||||
|
# llama-index-core
|
||||||
# openai
|
# openai
|
||||||
# pydantic
|
# pydantic
|
||||||
# pydantic-core
|
# pydantic-core
|
||||||
# pyee
|
# pyee
|
||||||
|
# sqlalchemy
|
||||||
# streamlit
|
# streamlit
|
||||||
|
# torch
|
||||||
|
# typing-inspect
|
||||||
|
typing-inspect==0.9.0
|
||||||
|
# via
|
||||||
|
# dataclasses-json
|
||||||
|
# llama-index-core
|
||||||
tzdata==2024.1
|
tzdata==2024.1
|
||||||
# via pandas
|
# via pandas
|
||||||
uritemplate==4.1.1
|
uritemplate==4.1.1
|
||||||
|
@ -330,6 +438,10 @@ watchdog==4.0.1
|
||||||
# via -r requirements.in
|
# via -r requirements.in
|
||||||
wcwidth==0.2.13
|
wcwidth==0.2.13
|
||||||
# via prompt-toolkit
|
# via prompt-toolkit
|
||||||
|
wrapt==1.16.0
|
||||||
|
# via
|
||||||
|
# deprecated
|
||||||
|
# llama-index-core
|
||||||
yarl==1.9.4
|
yarl==1.9.4
|
||||||
# via aiohttp
|
# via aiohttp
|
||||||
zipp==3.19.2
|
zipp==3.19.2
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue