mirror of
https://github.com/Significant-Gravitas/AutoGPT.git
synced 2026-01-21 04:57:58 -05:00
style(classic): update black to 24.10.0 and reformat
Update black version to match pre-commit hook (24.10.0) and reformat all files with the new version. Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
This commit is contained in:
@@ -182,5 +182,4 @@ class BaseChallenge(ABC):
|
||||
@abstractmethod
|
||||
async def evaluate_task_state(
|
||||
cls, agent: AgentApi, task_id: str
|
||||
) -> list[EvalResult]:
|
||||
...
|
||||
) -> list[EvalResult]: ...
|
||||
|
||||
@@ -68,9 +68,9 @@ class BuiltinChallengeSpec(BaseModel):
|
||||
class Eval(BaseModel):
|
||||
type: str
|
||||
scoring: Optional[Literal["percentage", "scale", "binary"]] = None
|
||||
template: Optional[
|
||||
Literal["rubric", "reference", "question", "custom"]
|
||||
] = None
|
||||
template: Optional[Literal["rubric", "reference", "question", "custom"]] = (
|
||||
None
|
||||
)
|
||||
examples: Optional[str] = None
|
||||
|
||||
@field_validator("scoring", "template")
|
||||
|
||||
@@ -83,13 +83,11 @@ def resolve_uri(uri: str) -> str:
|
||||
|
||||
class Eval(ABC):
|
||||
@abstractmethod
|
||||
def evaluate(self, string: str) -> bool:
|
||||
...
|
||||
def evaluate(self, string: str) -> bool: ...
|
||||
|
||||
@property
|
||||
@abstractmethod
|
||||
def description(self) -> str:
|
||||
...
|
||||
def description(self) -> str: ...
|
||||
|
||||
|
||||
class BaseStringEval(BaseModel, Eval):
|
||||
|
||||
@@ -142,24 +142,21 @@ class BaseAgent(Generic[AnyProposal], metaclass=AgentMeta):
|
||||
return self.config.send_token_limit or self.llm.max_tokens * 3 // 4
|
||||
|
||||
@abstractmethod
|
||||
async def propose_action(self) -> AnyProposal:
|
||||
...
|
||||
async def propose_action(self) -> AnyProposal: ...
|
||||
|
||||
@abstractmethod
|
||||
async def execute(
|
||||
self,
|
||||
proposal: AnyProposal,
|
||||
user_feedback: str = "",
|
||||
) -> ActionResult:
|
||||
...
|
||||
) -> ActionResult: ...
|
||||
|
||||
@abstractmethod
|
||||
async def do_not_execute(
|
||||
self,
|
||||
denied_proposal: AnyProposal,
|
||||
user_feedback: str,
|
||||
) -> ActionResult:
|
||||
...
|
||||
) -> ActionResult: ...
|
||||
|
||||
def reset_trace(self):
|
||||
self._trace = []
|
||||
@@ -167,8 +164,7 @@ class BaseAgent(Generic[AnyProposal], metaclass=AgentMeta):
|
||||
@overload
|
||||
async def run_pipeline(
|
||||
self, protocol_method: Callable[P, Iterator[T]], *args, retry_limit: int = 3
|
||||
) -> list[T]:
|
||||
...
|
||||
) -> list[T]: ...
|
||||
|
||||
@overload
|
||||
async def run_pipeline(
|
||||
@@ -176,8 +172,7 @@ class BaseAgent(Generic[AnyProposal], metaclass=AgentMeta):
|
||||
protocol_method: Callable[P, None | Awaitable[None]],
|
||||
*args,
|
||||
retry_limit: int = 3,
|
||||
) -> list[None]:
|
||||
...
|
||||
) -> list[None]: ...
|
||||
|
||||
async def run_pipeline(
|
||||
self,
|
||||
|
||||
@@ -23,29 +23,24 @@ class DirectiveProvider(AgentComponent):
|
||||
|
||||
class CommandProvider(AgentComponent):
|
||||
@abstractmethod
|
||||
def get_commands(self) -> Iterator["Command"]:
|
||||
...
|
||||
def get_commands(self) -> Iterator["Command"]: ...
|
||||
|
||||
|
||||
class MessageProvider(AgentComponent):
|
||||
@abstractmethod
|
||||
def get_messages(self) -> Iterator["ChatMessage"]:
|
||||
...
|
||||
def get_messages(self) -> Iterator["ChatMessage"]: ...
|
||||
|
||||
|
||||
class AfterParse(AgentComponent, Generic[AnyProposal]):
|
||||
@abstractmethod
|
||||
def after_parse(self, result: AnyProposal) -> None | Awaitable[None]:
|
||||
...
|
||||
def after_parse(self, result: AnyProposal) -> None | Awaitable[None]: ...
|
||||
|
||||
|
||||
class ExecutionFailure(AgentComponent):
|
||||
@abstractmethod
|
||||
def execution_failure(self, error: Exception) -> None | Awaitable[None]:
|
||||
...
|
||||
def execution_failure(self, error: Exception) -> None | Awaitable[None]: ...
|
||||
|
||||
|
||||
class AfterExecute(AgentComponent):
|
||||
@abstractmethod
|
||||
def after_execute(self, result: "ActionResult") -> None | Awaitable[None]:
|
||||
...
|
||||
def after_execute(self, result: "ActionResult") -> None | Awaitable[None]: ...
|
||||
|
||||
@@ -6,6 +6,7 @@ This module defines the API routes for the Agent service.
|
||||
Developers and contributors should be especially careful when making modifications
|
||||
to these routes to ensure consistency and correctness in the system's behavior.
|
||||
"""
|
||||
|
||||
import logging
|
||||
from typing import TYPE_CHECKING, Optional
|
||||
|
||||
|
||||
@@ -1,6 +1,7 @@
|
||||
"""
|
||||
This module contains configuration models and helpers for AutoGPT Forge.
|
||||
"""
|
||||
|
||||
from .ai_directives import AIDirectives
|
||||
from .ai_profile import AIProfile
|
||||
from .base import BaseConfig
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
"""HTML processing functions"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
from bs4 import BeautifulSoup
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
"""Text processing functions"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import logging
|
||||
|
||||
@@ -80,36 +80,30 @@ class GCSFileStorage(FileStorage):
|
||||
path: str | Path,
|
||||
mode: Literal["r", "w"] = "r",
|
||||
binary: Literal[False] = False,
|
||||
) -> TextIOWrapper:
|
||||
...
|
||||
) -> TextIOWrapper: ...
|
||||
|
||||
@overload
|
||||
def open_file(
|
||||
self, path: str | Path, mode: Literal["r"], binary: Literal[True]
|
||||
) -> BlobReader:
|
||||
...
|
||||
) -> BlobReader: ...
|
||||
|
||||
@overload
|
||||
def open_file(
|
||||
self, path: str | Path, mode: Literal["w"], binary: Literal[True]
|
||||
) -> BlobWriter:
|
||||
...
|
||||
) -> BlobWriter: ...
|
||||
|
||||
@overload
|
||||
def open_file(
|
||||
self, path: str | Path, mode: Literal["r", "w"], binary: Literal[True]
|
||||
) -> BlobWriter | BlobReader:
|
||||
...
|
||||
) -> BlobWriter | BlobReader: ...
|
||||
|
||||
@overload
|
||||
def open_file(self, path: str | Path, *, binary: Literal[True]) -> BlobReader:
|
||||
...
|
||||
def open_file(self, path: str | Path, *, binary: Literal[True]) -> BlobReader: ...
|
||||
|
||||
@overload
|
||||
def open_file(
|
||||
self, path: str | Path, mode: Literal["r", "w"] = "r", binary: bool = False
|
||||
) -> BlobReader | BlobWriter | TextIOWrapper:
|
||||
...
|
||||
) -> BlobReader | BlobWriter | TextIOWrapper: ...
|
||||
|
||||
# https://github.com/microsoft/pyright/issues/8007
|
||||
def open_file( # pyright: ignore[reportIncompatibleMethodOverride]
|
||||
|
||||
@@ -48,24 +48,20 @@ class LocalFileStorage(FileStorage):
|
||||
path: str | Path,
|
||||
mode: Literal["w", "r"] = "r",
|
||||
binary: Literal[False] = False,
|
||||
) -> TextIO:
|
||||
...
|
||||
) -> TextIO: ...
|
||||
|
||||
@overload
|
||||
def open_file(
|
||||
self, path: str | Path, mode: Literal["w", "r"], binary: Literal[True]
|
||||
) -> BinaryIO:
|
||||
...
|
||||
) -> BinaryIO: ...
|
||||
|
||||
@overload
|
||||
def open_file(self, path: str | Path, *, binary: Literal[True]) -> BinaryIO:
|
||||
...
|
||||
def open_file(self, path: str | Path, *, binary: Literal[True]) -> BinaryIO: ...
|
||||
|
||||
@overload
|
||||
def open_file(
|
||||
self, path: str | Path, mode: Literal["w", "r"] = "r", binary: bool = False
|
||||
) -> TextIO | BinaryIO:
|
||||
...
|
||||
) -> TextIO | BinaryIO: ...
|
||||
|
||||
def open_file(
|
||||
self, path: str | Path, mode: Literal["w", "r"] = "r", binary: bool = False
|
||||
|
||||
@@ -101,26 +101,22 @@ class S3FileStorage(FileStorage):
|
||||
path: str | Path,
|
||||
mode: Literal["r", "w"] = "r",
|
||||
binary: Literal[False] = False,
|
||||
) -> TextIOWrapper:
|
||||
...
|
||||
) -> TextIOWrapper: ...
|
||||
|
||||
@overload
|
||||
def open_file(
|
||||
self, path: str | Path, mode: Literal["r", "w"], binary: Literal[True]
|
||||
) -> S3BinaryIOWrapper:
|
||||
...
|
||||
) -> S3BinaryIOWrapper: ...
|
||||
|
||||
@overload
|
||||
def open_file(
|
||||
self, path: str | Path, *, binary: Literal[True]
|
||||
) -> S3BinaryIOWrapper:
|
||||
...
|
||||
) -> S3BinaryIOWrapper: ...
|
||||
|
||||
@overload
|
||||
def open_file(
|
||||
self, path: str | Path, mode: Literal["r", "w"] = "r", binary: bool = False
|
||||
) -> S3BinaryIOWrapper | TextIOWrapper:
|
||||
...
|
||||
) -> S3BinaryIOWrapper | TextIOWrapper: ...
|
||||
|
||||
def open_file(
|
||||
self, path: str | Path, mode: Literal["r", "w"] = "r", binary: bool = False
|
||||
|
||||
@@ -10,13 +10,10 @@ from .schema import ChatPrompt, LanguageModelClassification
|
||||
class PromptStrategy(abc.ABC):
|
||||
@property
|
||||
@abc.abstractmethod
|
||||
def llm_classification(self) -> LanguageModelClassification:
|
||||
...
|
||||
def llm_classification(self) -> LanguageModelClassification: ...
|
||||
|
||||
@abc.abstractmethod
|
||||
def build_prompt(self, *_, **kwargs) -> ChatPrompt:
|
||||
...
|
||||
def build_prompt(self, *_, **kwargs) -> ChatPrompt: ...
|
||||
|
||||
@abc.abstractmethod
|
||||
def parse_response_content(self, response: "AssistantChatMessage") -> Any:
|
||||
...
|
||||
def parse_response_content(self, response: "AssistantChatMessage") -> Any: ...
|
||||
|
||||
@@ -309,20 +309,16 @@ class BaseModelProvider(
|
||||
@abc.abstractmethod
|
||||
async def get_available_models(
|
||||
self,
|
||||
) -> Sequence["ChatModelInfo[_ModelName] | EmbeddingModelInfo[_ModelName]"]:
|
||||
...
|
||||
) -> Sequence["ChatModelInfo[_ModelName] | EmbeddingModelInfo[_ModelName]"]: ...
|
||||
|
||||
@abc.abstractmethod
|
||||
def count_tokens(self, text: str, model_name: _ModelName) -> int:
|
||||
...
|
||||
def count_tokens(self, text: str, model_name: _ModelName) -> int: ...
|
||||
|
||||
@abc.abstractmethod
|
||||
def get_tokenizer(self, model_name: _ModelName) -> "ModelTokenizer[Any]":
|
||||
...
|
||||
def get_tokenizer(self, model_name: _ModelName) -> "ModelTokenizer[Any]": ...
|
||||
|
||||
@abc.abstractmethod
|
||||
def get_token_limit(self, model_name: _ModelName) -> int:
|
||||
...
|
||||
def get_token_limit(self, model_name: _ModelName) -> int: ...
|
||||
|
||||
def get_incurred_cost(self) -> float:
|
||||
if self._budget:
|
||||
@@ -339,12 +335,10 @@ class ModelTokenizer(Protocol, Generic[_T]):
|
||||
"""A ModelTokenizer provides tokenization specific to a model."""
|
||||
|
||||
@abc.abstractmethod
|
||||
def encode(self, text: str) -> list[_T]:
|
||||
...
|
||||
def encode(self, text: str) -> list[_T]: ...
|
||||
|
||||
@abc.abstractmethod
|
||||
def decode(self, tokens: list[_T]) -> str:
|
||||
...
|
||||
def decode(self, tokens: list[_T]) -> str: ...
|
||||
|
||||
|
||||
####################
|
||||
@@ -371,8 +365,7 @@ class BaseEmbeddingModelProvider(BaseModelProvider[_ModelName, _ModelProviderSet
|
||||
@abc.abstractmethod
|
||||
async def get_available_embedding_models(
|
||||
self,
|
||||
) -> Sequence[EmbeddingModelInfo[_ModelName]]:
|
||||
...
|
||||
) -> Sequence[EmbeddingModelInfo[_ModelName]]: ...
|
||||
|
||||
@abc.abstractmethod
|
||||
async def create_embedding(
|
||||
@@ -381,8 +374,7 @@ class BaseEmbeddingModelProvider(BaseModelProvider[_ModelName, _ModelProviderSet
|
||||
model_name: _ModelName,
|
||||
embedding_parser: Callable[[Embedding], Embedding],
|
||||
**kwargs,
|
||||
) -> EmbeddingModelResponse:
|
||||
...
|
||||
) -> EmbeddingModelResponse: ...
|
||||
|
||||
|
||||
###############
|
||||
@@ -413,16 +405,14 @@ class BaseChatModelProvider(BaseModelProvider[_ModelName, _ModelProviderSettings
|
||||
@abc.abstractmethod
|
||||
async def get_available_chat_models(
|
||||
self,
|
||||
) -> Sequence[ChatModelInfo[_ModelName]]:
|
||||
...
|
||||
) -> Sequence[ChatModelInfo[_ModelName]]: ...
|
||||
|
||||
@abc.abstractmethod
|
||||
def count_message_tokens(
|
||||
self,
|
||||
messages: ChatMessage | list[ChatMessage],
|
||||
model_name: _ModelName,
|
||||
) -> int:
|
||||
...
|
||||
) -> int: ...
|
||||
|
||||
@abc.abstractmethod
|
||||
async def create_chat_completion(
|
||||
@@ -434,5 +424,4 @@ class BaseChatModelProvider(BaseModelProvider[_ModelName, _ModelProviderSettings
|
||||
max_output_tokens: Optional[int] = None,
|
||||
prefill_response: str = "",
|
||||
**kwargs,
|
||||
) -> ChatModelResponse[_T]:
|
||||
...
|
||||
) -> ChatModelResponse[_T]: ...
|
||||
|
||||
@@ -47,9 +47,11 @@ def validate_tool_calls(
|
||||
is_valid, validation_errors = function.validate_call(function_call)
|
||||
if not is_valid:
|
||||
fmt_errors = [
|
||||
f"{'.'.join(str(p) for p in f.path)}: {f.message}"
|
||||
if f.path
|
||||
else f.message
|
||||
(
|
||||
f"{'.'.join(str(p) for p in f.path)}: {f.message}"
|
||||
if f.path
|
||||
else f.message
|
||||
)
|
||||
for f in validation_errors
|
||||
]
|
||||
errors.append(
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
"""Logging module for Auto-GPT."""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import enum
|
||||
|
||||
@@ -155,13 +155,11 @@ class JSONSchema(BaseModel):
|
||||
|
||||
|
||||
@overload
|
||||
def _resolve_type_refs_in_schema(schema: dict, definitions: dict) -> dict:
|
||||
...
|
||||
def _resolve_type_refs_in_schema(schema: dict, definitions: dict) -> dict: ...
|
||||
|
||||
|
||||
@overload
|
||||
def _resolve_type_refs_in_schema(schema: list, definitions: dict) -> list:
|
||||
...
|
||||
def _resolve_type_refs_in_schema(schema: list, definitions: dict) -> list: ...
|
||||
|
||||
|
||||
def _resolve_type_refs_in_schema(schema: dict | list, definitions: dict) -> dict | list:
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
"""This module contains the (speech recognition and) speech synthesis functions."""
|
||||
|
||||
from .say import TextToSpeechProvider, TTSConfig
|
||||
|
||||
__all__ = ["TextToSpeechProvider", "TTSConfig"]
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
"""Base class for all voice classes."""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import abc
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
"""ElevenLabs speech module"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import logging
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
""" GTTS Voice. """
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import os
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
""" MacOS TTS Voice. """
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import subprocess
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
""" Text to speech module """
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import os
|
||||
@@ -31,11 +32,11 @@ class TTSConfig(SystemConfiguration):
|
||||
or (
|
||||
"macos"
|
||||
if os.getenv("USE_MAC_OS_TTS")
|
||||
else "elevenlabs"
|
||||
if os.getenv("ELEVENLABS_API_KEY")
|
||||
else "streamelements"
|
||||
if os.getenv("USE_BRIAN_TTS")
|
||||
else "gtts"
|
||||
else (
|
||||
"elevenlabs"
|
||||
if os.getenv("ELEVENLABS_API_KEY")
|
||||
else "streamelements" if os.getenv("USE_BRIAN_TTS") else "gtts"
|
||||
)
|
||||
),
|
||||
) # type: ignore
|
||||
|
||||
|
||||
@@ -15,8 +15,7 @@ logger = logging.getLogger(__name__)
|
||||
|
||||
class ParserStrategy(ABC):
|
||||
@abstractmethod
|
||||
def read(self, file: BinaryIO) -> str:
|
||||
...
|
||||
def read(self, file: BinaryIO) -> str: ...
|
||||
|
||||
|
||||
# Basic text file reading
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
"""AutoGPT: A GPT powered AI Assistant"""
|
||||
|
||||
import autogpt.app.cli
|
||||
|
||||
if __name__ == "__main__":
|
||||
|
||||
@@ -885,9 +885,7 @@ def print_assistant_thoughts(
|
||||
thoughts_text = remove_ansi_escape(
|
||||
thoughts.text
|
||||
if isinstance(thoughts, AssistantThoughts)
|
||||
else thoughts.summary()
|
||||
if isinstance(thoughts, ModelWithSummary)
|
||||
else thoughts
|
||||
else thoughts.summary() if isinstance(thoughts, ModelWithSummary) else thoughts
|
||||
)
|
||||
print_attribute(
|
||||
f"{ai_name.upper()} THOUGHTS", thoughts_text, title_color=Fore.YELLOW
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
"""A simple spinner module"""
|
||||
|
||||
import itertools
|
||||
import sys
|
||||
import threading
|
||||
|
||||
54
classic/poetry.lock
generated
54
classic/poetry.lock
generated
@@ -382,34 +382,34 @@ lxml = ["lxml"]
|
||||
|
||||
[[package]]
|
||||
name = "black"
|
||||
version = "23.12.1"
|
||||
version = "24.10.0"
|
||||
description = "The uncompromising code formatter."
|
||||
optional = false
|
||||
python-versions = ">=3.8"
|
||||
python-versions = ">=3.9"
|
||||
groups = ["dev"]
|
||||
files = [
|
||||
{file = "black-23.12.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0aaf6041986767a5e0ce663c7a2f0e9eaf21e6ff87a5f95cbf3675bfd4c41d2"},
|
||||
{file = "black-23.12.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c88b3711d12905b74206227109272673edce0cb29f27e1385f33b0163c414bba"},
|
||||
{file = "black-23.12.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a920b569dc6b3472513ba6ddea21f440d4b4c699494d2e972a1753cdc25df7b0"},
|
||||
{file = "black-23.12.1-cp310-cp310-win_amd64.whl", hash = "sha256:3fa4be75ef2a6b96ea8d92b1587dd8cb3a35c7e3d51f0738ced0781c3aa3a5a3"},
|
||||
{file = "black-23.12.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:8d4df77958a622f9b5a4c96edb4b8c0034f8434032ab11077ec6c56ae9f384ba"},
|
||||
{file = "black-23.12.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:602cfb1196dc692424c70b6507593a2b29aac0547c1be9a1d1365f0d964c353b"},
|
||||
{file = "black-23.12.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9c4352800f14be5b4864016882cdba10755bd50805c95f728011bcb47a4afd59"},
|
||||
{file = "black-23.12.1-cp311-cp311-win_amd64.whl", hash = "sha256:0808494f2b2df923ffc5723ed3c7b096bd76341f6213989759287611e9837d50"},
|
||||
{file = "black-23.12.1-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:25e57fd232a6d6ff3f4478a6fd0580838e47c93c83eaf1ccc92d4faf27112c4e"},
|
||||
{file = "black-23.12.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:2d9e13db441c509a3763a7a3d9a49ccc1b4e974a47be4e08ade2a228876500ec"},
|
||||
{file = "black-23.12.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6d1bd9c210f8b109b1762ec9fd36592fdd528485aadb3f5849b2740ef17e674e"},
|
||||
{file = "black-23.12.1-cp312-cp312-win_amd64.whl", hash = "sha256:ae76c22bde5cbb6bfd211ec343ded2163bba7883c7bc77f6b756a1049436fbb9"},
|
||||
{file = "black-23.12.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1fa88a0f74e50e4487477bc0bb900c6781dbddfdfa32691e780bf854c3b4a47f"},
|
||||
{file = "black-23.12.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:a4d6a9668e45ad99d2f8ec70d5c8c04ef4f32f648ef39048d010b0689832ec6d"},
|
||||
{file = "black-23.12.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b18fb2ae6c4bb63eebe5be6bd869ba2f14fd0259bda7d18a46b764d8fb86298a"},
|
||||
{file = "black-23.12.1-cp38-cp38-win_amd64.whl", hash = "sha256:c04b6d9d20e9c13f43eee8ea87d44156b8505ca8a3c878773f68b4e4812a421e"},
|
||||
{file = "black-23.12.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3e1b38b3135fd4c025c28c55ddfc236b05af657828a8a6abe5deec419a0b7055"},
|
||||
{file = "black-23.12.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4f0031eaa7b921db76decd73636ef3a12c942ed367d8c3841a0739412b260a54"},
|
||||
{file = "black-23.12.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:97e56155c6b737854e60a9ab1c598ff2533d57e7506d97af5481141671abf3ea"},
|
||||
{file = "black-23.12.1-cp39-cp39-win_amd64.whl", hash = "sha256:dd15245c8b68fe2b6bd0f32c1556509d11bb33aec9b5d0866dd8e2ed3dba09c2"},
|
||||
{file = "black-23.12.1-py3-none-any.whl", hash = "sha256:78baad24af0f033958cad29731e27363183e140962595def56423e626f4bee3e"},
|
||||
{file = "black-23.12.1.tar.gz", hash = "sha256:4ce3ef14ebe8d9509188014d96af1c456a910d5b5cbf434a09fef7e024b3d0d5"},
|
||||
{file = "black-24.10.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e6668650ea4b685440857138e5fe40cde4d652633b1bdffc62933d0db4ed9812"},
|
||||
{file = "black-24.10.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1c536fcf674217e87b8cc3657b81809d3c085d7bf3ef262ead700da345bfa6ea"},
|
||||
{file = "black-24.10.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:649fff99a20bd06c6f727d2a27f401331dc0cc861fb69cde910fe95b01b5928f"},
|
||||
{file = "black-24.10.0-cp310-cp310-win_amd64.whl", hash = "sha256:fe4d6476887de70546212c99ac9bd803d90b42fc4767f058a0baa895013fbb3e"},
|
||||
{file = "black-24.10.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5a2221696a8224e335c28816a9d331a6c2ae15a2ee34ec857dcf3e45dbfa99ad"},
|
||||
{file = "black-24.10.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f9da3333530dbcecc1be13e69c250ed8dfa67f43c4005fb537bb426e19200d50"},
|
||||
{file = "black-24.10.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4007b1393d902b48b36958a216c20c4482f601569d19ed1df294a496eb366392"},
|
||||
{file = "black-24.10.0-cp311-cp311-win_amd64.whl", hash = "sha256:394d4ddc64782e51153eadcaaca95144ac4c35e27ef9b0a42e121ae7e57a9175"},
|
||||
{file = "black-24.10.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:b5e39e0fae001df40f95bd8cc36b9165c5e2ea88900167bddf258bacef9bbdc3"},
|
||||
{file = "black-24.10.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:d37d422772111794b26757c5b55a3eade028aa3fde43121ab7b673d050949d65"},
|
||||
{file = "black-24.10.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:14b3502784f09ce2443830e3133dacf2c0110d45191ed470ecb04d0f5f6fcb0f"},
|
||||
{file = "black-24.10.0-cp312-cp312-win_amd64.whl", hash = "sha256:30d2c30dc5139211dda799758559d1b049f7f14c580c409d6ad925b74a4208a8"},
|
||||
{file = "black-24.10.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:1cbacacb19e922a1d75ef2b6ccaefcd6e93a2c05ede32f06a21386a04cedb981"},
|
||||
{file = "black-24.10.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:1f93102e0c5bb3907451063e08b9876dbeac810e7da5a8bfb7aeb5a9ef89066b"},
|
||||
{file = "black-24.10.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ddacb691cdcdf77b96f549cf9591701d8db36b2f19519373d60d31746068dbf2"},
|
||||
{file = "black-24.10.0-cp313-cp313-win_amd64.whl", hash = "sha256:680359d932801c76d2e9c9068d05c6b107f2584b2a5b88831c83962eb9984c1b"},
|
||||
{file = "black-24.10.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:17374989640fbca88b6a448129cd1745c5eb8d9547b464f281b251dd00155ccd"},
|
||||
{file = "black-24.10.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:63f626344343083322233f175aaf372d326de8436f5928c042639a4afbbf1d3f"},
|
||||
{file = "black-24.10.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ccfa1d0cb6200857f1923b602f978386a3a2758a65b52e0950299ea014be6800"},
|
||||
{file = "black-24.10.0-cp39-cp39-win_amd64.whl", hash = "sha256:2cd9c95431d94adc56600710f8813ee27eea544dd118d45896bb734e9d7a0dc7"},
|
||||
{file = "black-24.10.0-py3-none-any.whl", hash = "sha256:3bb2b7a1f7b685f85b11fed1ef10f8a9148bceb49853e47a294a3dd963c1dd7d"},
|
||||
{file = "black-24.10.0.tar.gz", hash = "sha256:846ea64c97afe3bc677b761787993be4991810ecc7a4a937816dd6bddedc4875"},
|
||||
]
|
||||
|
||||
[package.dependencies]
|
||||
@@ -421,7 +421,7 @@ platformdirs = ">=2"
|
||||
|
||||
[package.extras]
|
||||
colorama = ["colorama (>=0.4.3)"]
|
||||
d = ["aiohttp (>=3.7.4) ; sys_platform != \"win32\" or implementation_name != \"pypy\"", "aiohttp (>=3.7.4,!=3.9.0) ; sys_platform == \"win32\" and implementation_name == \"pypy\""]
|
||||
d = ["aiohttp (>=3.10)"]
|
||||
jupyter = ["ipython (>=7.8.0)", "tokenize-rt (>=3.2.0)"]
|
||||
uvloop = ["uvloop (>=0.15.2)"]
|
||||
|
||||
@@ -2984,7 +2984,7 @@ description = "Fast transfer of large files with the Hugging Face Hub."
|
||||
optional = false
|
||||
python-versions = ">=3.8"
|
||||
groups = ["main"]
|
||||
markers = "(sys_platform == \"linux\" or platform_machine == \"x86_64\" or platform_machine == \"amd64\" or platform_machine == \"AMD64\" or platform_machine == \"arm64\" or platform_machine == \"aarch64\") and (platform_machine == \"x86_64\" or platform_machine == \"aarch64\" or platform_machine == \"amd64\" or platform_machine == \"AMD64\" or platform_machine == \"arm64\")"
|
||||
markers = "platform_machine == \"x86_64\" or platform_machine == \"amd64\" or platform_machine == \"AMD64\" or platform_machine == \"arm64\" or platform_machine == \"aarch64\""
|
||||
files = [
|
||||
{file = "hf_xet-1.2.0-cp313-cp313t-macosx_10_12_x86_64.whl", hash = "sha256:ceeefcd1b7aed4956ae8499e2199607765fbd1c60510752003b6cc0b8413b649"},
|
||||
{file = "hf_xet-1.2.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:b70218dd548e9840224df5638fdc94bd033552963cfa97f9170829381179c813"},
|
||||
@@ -8774,4 +8774,4 @@ type = ["pytest-mypy"]
|
||||
[metadata]
|
||||
lock-version = "2.1"
|
||||
python-versions = "^3.12"
|
||||
content-hash = "bd647e9d41cdc55cb4b3d784746e3e09378180b730e3de3bf3e2df1899c113bd"
|
||||
content-hash = "9d632f4341ef2e49aa81978032caf9605fd258aff8176263ccb1766b0a1ef4b1"
|
||||
|
||||
@@ -111,7 +111,7 @@ pytest = "^8.0"
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
# Formatting & linting
|
||||
black = "^23.12.1"
|
||||
black = "^24.10.0"
|
||||
flake8 = "^7.0.0"
|
||||
isort = "^5.13.1"
|
||||
pre-commit = "^3.3.3"
|
||||
|
||||
Reference in New Issue
Block a user