style(classic): update black to 24.10.0 and reformat

Update black version to match pre-commit hook (24.10.0) and reformat
all files with the new version.

Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
This commit is contained in:
Nicholas Tindle
2026-01-20 10:51:54 -06:00
parent 5e22a1888a
commit 326554d89a
29 changed files with 99 additions and 130 deletions

View File

@@ -182,5 +182,4 @@ class BaseChallenge(ABC):
@abstractmethod
async def evaluate_task_state(
cls, agent: AgentApi, task_id: str
) -> list[EvalResult]:
...
) -> list[EvalResult]: ...

View File

@@ -68,9 +68,9 @@ class BuiltinChallengeSpec(BaseModel):
class Eval(BaseModel):
type: str
scoring: Optional[Literal["percentage", "scale", "binary"]] = None
template: Optional[
Literal["rubric", "reference", "question", "custom"]
] = None
template: Optional[Literal["rubric", "reference", "question", "custom"]] = (
None
)
examples: Optional[str] = None
@field_validator("scoring", "template")

View File

@@ -83,13 +83,11 @@ def resolve_uri(uri: str) -> str:
class Eval(ABC):
@abstractmethod
def evaluate(self, string: str) -> bool:
...
def evaluate(self, string: str) -> bool: ...
@property
@abstractmethod
def description(self) -> str:
...
def description(self) -> str: ...
class BaseStringEval(BaseModel, Eval):

View File

@@ -142,24 +142,21 @@ class BaseAgent(Generic[AnyProposal], metaclass=AgentMeta):
return self.config.send_token_limit or self.llm.max_tokens * 3 // 4
@abstractmethod
async def propose_action(self) -> AnyProposal:
...
async def propose_action(self) -> AnyProposal: ...
@abstractmethod
async def execute(
self,
proposal: AnyProposal,
user_feedback: str = "",
) -> ActionResult:
...
) -> ActionResult: ...
@abstractmethod
async def do_not_execute(
self,
denied_proposal: AnyProposal,
user_feedback: str,
) -> ActionResult:
...
) -> ActionResult: ...
def reset_trace(self):
self._trace = []
@@ -167,8 +164,7 @@ class BaseAgent(Generic[AnyProposal], metaclass=AgentMeta):
@overload
async def run_pipeline(
self, protocol_method: Callable[P, Iterator[T]], *args, retry_limit: int = 3
) -> list[T]:
...
) -> list[T]: ...
@overload
async def run_pipeline(
@@ -176,8 +172,7 @@ class BaseAgent(Generic[AnyProposal], metaclass=AgentMeta):
protocol_method: Callable[P, None | Awaitable[None]],
*args,
retry_limit: int = 3,
) -> list[None]:
...
) -> list[None]: ...
async def run_pipeline(
self,

View File

@@ -23,29 +23,24 @@ class DirectiveProvider(AgentComponent):
class CommandProvider(AgentComponent):
@abstractmethod
def get_commands(self) -> Iterator["Command"]:
...
def get_commands(self) -> Iterator["Command"]: ...
class MessageProvider(AgentComponent):
@abstractmethod
def get_messages(self) -> Iterator["ChatMessage"]:
...
def get_messages(self) -> Iterator["ChatMessage"]: ...
class AfterParse(AgentComponent, Generic[AnyProposal]):
@abstractmethod
def after_parse(self, result: AnyProposal) -> None | Awaitable[None]:
...
def after_parse(self, result: AnyProposal) -> None | Awaitable[None]: ...
class ExecutionFailure(AgentComponent):
@abstractmethod
def execution_failure(self, error: Exception) -> None | Awaitable[None]:
...
def execution_failure(self, error: Exception) -> None | Awaitable[None]: ...
class AfterExecute(AgentComponent):
@abstractmethod
def after_execute(self, result: "ActionResult") -> None | Awaitable[None]:
...
def after_execute(self, result: "ActionResult") -> None | Awaitable[None]: ...

View File

@@ -6,6 +6,7 @@ This module defines the API routes for the Agent service.
Developers and contributors should be especially careful when making modifications
to these routes to ensure consistency and correctness in the system's behavior.
"""
import logging
from typing import TYPE_CHECKING, Optional

View File

@@ -1,6 +1,7 @@
"""
This module contains configuration models and helpers for AutoGPT Forge.
"""
from .ai_directives import AIDirectives
from .ai_profile import AIProfile
from .base import BaseConfig

View File

@@ -1,4 +1,5 @@
"""HTML processing functions"""
from __future__ import annotations
from bs4 import BeautifulSoup

View File

@@ -1,4 +1,5 @@
"""Text processing functions"""
from __future__ import annotations
import logging

View File

@@ -80,36 +80,30 @@ class GCSFileStorage(FileStorage):
path: str | Path,
mode: Literal["r", "w"] = "r",
binary: Literal[False] = False,
) -> TextIOWrapper:
...
) -> TextIOWrapper: ...
@overload
def open_file(
self, path: str | Path, mode: Literal["r"], binary: Literal[True]
) -> BlobReader:
...
) -> BlobReader: ...
@overload
def open_file(
self, path: str | Path, mode: Literal["w"], binary: Literal[True]
) -> BlobWriter:
...
) -> BlobWriter: ...
@overload
def open_file(
self, path: str | Path, mode: Literal["r", "w"], binary: Literal[True]
) -> BlobWriter | BlobReader:
...
) -> BlobWriter | BlobReader: ...
@overload
def open_file(self, path: str | Path, *, binary: Literal[True]) -> BlobReader:
...
def open_file(self, path: str | Path, *, binary: Literal[True]) -> BlobReader: ...
@overload
def open_file(
self, path: str | Path, mode: Literal["r", "w"] = "r", binary: bool = False
) -> BlobReader | BlobWriter | TextIOWrapper:
...
) -> BlobReader | BlobWriter | TextIOWrapper: ...
# https://github.com/microsoft/pyright/issues/8007
def open_file( # pyright: ignore[reportIncompatibleMethodOverride]

View File

@@ -48,24 +48,20 @@ class LocalFileStorage(FileStorage):
path: str | Path,
mode: Literal["w", "r"] = "r",
binary: Literal[False] = False,
) -> TextIO:
...
) -> TextIO: ...
@overload
def open_file(
self, path: str | Path, mode: Literal["w", "r"], binary: Literal[True]
) -> BinaryIO:
...
) -> BinaryIO: ...
@overload
def open_file(self, path: str | Path, *, binary: Literal[True]) -> BinaryIO:
...
def open_file(self, path: str | Path, *, binary: Literal[True]) -> BinaryIO: ...
@overload
def open_file(
self, path: str | Path, mode: Literal["w", "r"] = "r", binary: bool = False
) -> TextIO | BinaryIO:
...
) -> TextIO | BinaryIO: ...
def open_file(
self, path: str | Path, mode: Literal["w", "r"] = "r", binary: bool = False

View File

@@ -101,26 +101,22 @@ class S3FileStorage(FileStorage):
path: str | Path,
mode: Literal["r", "w"] = "r",
binary: Literal[False] = False,
) -> TextIOWrapper:
...
) -> TextIOWrapper: ...
@overload
def open_file(
self, path: str | Path, mode: Literal["r", "w"], binary: Literal[True]
) -> S3BinaryIOWrapper:
...
) -> S3BinaryIOWrapper: ...
@overload
def open_file(
self, path: str | Path, *, binary: Literal[True]
) -> S3BinaryIOWrapper:
...
) -> S3BinaryIOWrapper: ...
@overload
def open_file(
self, path: str | Path, mode: Literal["r", "w"] = "r", binary: bool = False
) -> S3BinaryIOWrapper | TextIOWrapper:
...
) -> S3BinaryIOWrapper | TextIOWrapper: ...
def open_file(
self, path: str | Path, mode: Literal["r", "w"] = "r", binary: bool = False

View File

@@ -10,13 +10,10 @@ from .schema import ChatPrompt, LanguageModelClassification
class PromptStrategy(abc.ABC):
@property
@abc.abstractmethod
def llm_classification(self) -> LanguageModelClassification:
...
def llm_classification(self) -> LanguageModelClassification: ...
@abc.abstractmethod
def build_prompt(self, *_, **kwargs) -> ChatPrompt:
...
def build_prompt(self, *_, **kwargs) -> ChatPrompt: ...
@abc.abstractmethod
def parse_response_content(self, response: "AssistantChatMessage") -> Any:
...
def parse_response_content(self, response: "AssistantChatMessage") -> Any: ...

View File

@@ -309,20 +309,16 @@ class BaseModelProvider(
@abc.abstractmethod
async def get_available_models(
self,
) -> Sequence["ChatModelInfo[_ModelName] | EmbeddingModelInfo[_ModelName]"]:
...
) -> Sequence["ChatModelInfo[_ModelName] | EmbeddingModelInfo[_ModelName]"]: ...
@abc.abstractmethod
def count_tokens(self, text: str, model_name: _ModelName) -> int:
...
def count_tokens(self, text: str, model_name: _ModelName) -> int: ...
@abc.abstractmethod
def get_tokenizer(self, model_name: _ModelName) -> "ModelTokenizer[Any]":
...
def get_tokenizer(self, model_name: _ModelName) -> "ModelTokenizer[Any]": ...
@abc.abstractmethod
def get_token_limit(self, model_name: _ModelName) -> int:
...
def get_token_limit(self, model_name: _ModelName) -> int: ...
def get_incurred_cost(self) -> float:
if self._budget:
@@ -339,12 +335,10 @@ class ModelTokenizer(Protocol, Generic[_T]):
"""A ModelTokenizer provides tokenization specific to a model."""
@abc.abstractmethod
def encode(self, text: str) -> list[_T]:
...
def encode(self, text: str) -> list[_T]: ...
@abc.abstractmethod
def decode(self, tokens: list[_T]) -> str:
...
def decode(self, tokens: list[_T]) -> str: ...
####################
@@ -371,8 +365,7 @@ class BaseEmbeddingModelProvider(BaseModelProvider[_ModelName, _ModelProviderSet
@abc.abstractmethod
async def get_available_embedding_models(
self,
) -> Sequence[EmbeddingModelInfo[_ModelName]]:
...
) -> Sequence[EmbeddingModelInfo[_ModelName]]: ...
@abc.abstractmethod
async def create_embedding(
@@ -381,8 +374,7 @@ class BaseEmbeddingModelProvider(BaseModelProvider[_ModelName, _ModelProviderSet
model_name: _ModelName,
embedding_parser: Callable[[Embedding], Embedding],
**kwargs,
) -> EmbeddingModelResponse:
...
) -> EmbeddingModelResponse: ...
###############
@@ -413,16 +405,14 @@ class BaseChatModelProvider(BaseModelProvider[_ModelName, _ModelProviderSettings
@abc.abstractmethod
async def get_available_chat_models(
self,
) -> Sequence[ChatModelInfo[_ModelName]]:
...
) -> Sequence[ChatModelInfo[_ModelName]]: ...
@abc.abstractmethod
def count_message_tokens(
self,
messages: ChatMessage | list[ChatMessage],
model_name: _ModelName,
) -> int:
...
) -> int: ...
@abc.abstractmethod
async def create_chat_completion(
@@ -434,5 +424,4 @@ class BaseChatModelProvider(BaseModelProvider[_ModelName, _ModelProviderSettings
max_output_tokens: Optional[int] = None,
prefill_response: str = "",
**kwargs,
) -> ChatModelResponse[_T]:
...
) -> ChatModelResponse[_T]: ...

View File

@@ -47,9 +47,11 @@ def validate_tool_calls(
is_valid, validation_errors = function.validate_call(function_call)
if not is_valid:
fmt_errors = [
f"{'.'.join(str(p) for p in f.path)}: {f.message}"
if f.path
else f.message
(
f"{'.'.join(str(p) for p in f.path)}: {f.message}"
if f.path
else f.message
)
for f in validation_errors
]
errors.append(

View File

@@ -1,4 +1,5 @@
"""Logging module for Auto-GPT."""
from __future__ import annotations
import enum

View File

@@ -155,13 +155,11 @@ class JSONSchema(BaseModel):
@overload
def _resolve_type_refs_in_schema(schema: dict, definitions: dict) -> dict:
...
def _resolve_type_refs_in_schema(schema: dict, definitions: dict) -> dict: ...
@overload
def _resolve_type_refs_in_schema(schema: list, definitions: dict) -> list:
...
def _resolve_type_refs_in_schema(schema: list, definitions: dict) -> list: ...
def _resolve_type_refs_in_schema(schema: dict | list, definitions: dict) -> dict | list:

View File

@@ -1,4 +1,5 @@
"""This module contains the (speech recognition and) speech synthesis functions."""
from .say import TextToSpeechProvider, TTSConfig
__all__ = ["TextToSpeechProvider", "TTSConfig"]

View File

@@ -1,4 +1,5 @@
"""Base class for all voice classes."""
from __future__ import annotations
import abc

View File

@@ -1,4 +1,5 @@
"""ElevenLabs speech module"""
from __future__ import annotations
import logging

View File

@@ -1,4 +1,5 @@
""" GTTS Voice. """
from __future__ import annotations
import os

View File

@@ -1,4 +1,5 @@
""" MacOS TTS Voice. """
from __future__ import annotations
import subprocess

View File

@@ -1,4 +1,5 @@
""" Text to speech module """
from __future__ import annotations
import os
@@ -31,11 +32,11 @@ class TTSConfig(SystemConfiguration):
or (
"macos"
if os.getenv("USE_MAC_OS_TTS")
else "elevenlabs"
if os.getenv("ELEVENLABS_API_KEY")
else "streamelements"
if os.getenv("USE_BRIAN_TTS")
else "gtts"
else (
"elevenlabs"
if os.getenv("ELEVENLABS_API_KEY")
else "streamelements" if os.getenv("USE_BRIAN_TTS") else "gtts"
)
),
) # type: ignore

View File

@@ -15,8 +15,7 @@ logger = logging.getLogger(__name__)
class ParserStrategy(ABC):
@abstractmethod
def read(self, file: BinaryIO) -> str:
...
def read(self, file: BinaryIO) -> str: ...
# Basic text file reading

View File

@@ -1,4 +1,5 @@
"""AutoGPT: A GPT powered AI Assistant"""
import autogpt.app.cli
if __name__ == "__main__":

View File

@@ -885,9 +885,7 @@ def print_assistant_thoughts(
thoughts_text = remove_ansi_escape(
thoughts.text
if isinstance(thoughts, AssistantThoughts)
else thoughts.summary()
if isinstance(thoughts, ModelWithSummary)
else thoughts
else thoughts.summary() if isinstance(thoughts, ModelWithSummary) else thoughts
)
print_attribute(
f"{ai_name.upper()} THOUGHTS", thoughts_text, title_color=Fore.YELLOW

View File

@@ -1,4 +1,5 @@
"""A simple spinner module"""
import itertools
import sys
import threading

54
classic/poetry.lock generated
View File

@@ -382,34 +382,34 @@ lxml = ["lxml"]
[[package]]
name = "black"
version = "23.12.1"
version = "24.10.0"
description = "The uncompromising code formatter."
optional = false
python-versions = ">=3.8"
python-versions = ">=3.9"
groups = ["dev"]
files = [
{file = "black-23.12.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0aaf6041986767a5e0ce663c7a2f0e9eaf21e6ff87a5f95cbf3675bfd4c41d2"},
{file = "black-23.12.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c88b3711d12905b74206227109272673edce0cb29f27e1385f33b0163c414bba"},
{file = "black-23.12.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a920b569dc6b3472513ba6ddea21f440d4b4c699494d2e972a1753cdc25df7b0"},
{file = "black-23.12.1-cp310-cp310-win_amd64.whl", hash = "sha256:3fa4be75ef2a6b96ea8d92b1587dd8cb3a35c7e3d51f0738ced0781c3aa3a5a3"},
{file = "black-23.12.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:8d4df77958a622f9b5a4c96edb4b8c0034f8434032ab11077ec6c56ae9f384ba"},
{file = "black-23.12.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:602cfb1196dc692424c70b6507593a2b29aac0547c1be9a1d1365f0d964c353b"},
{file = "black-23.12.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9c4352800f14be5b4864016882cdba10755bd50805c95f728011bcb47a4afd59"},
{file = "black-23.12.1-cp311-cp311-win_amd64.whl", hash = "sha256:0808494f2b2df923ffc5723ed3c7b096bd76341f6213989759287611e9837d50"},
{file = "black-23.12.1-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:25e57fd232a6d6ff3f4478a6fd0580838e47c93c83eaf1ccc92d4faf27112c4e"},
{file = "black-23.12.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:2d9e13db441c509a3763a7a3d9a49ccc1b4e974a47be4e08ade2a228876500ec"},
{file = "black-23.12.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6d1bd9c210f8b109b1762ec9fd36592fdd528485aadb3f5849b2740ef17e674e"},
{file = "black-23.12.1-cp312-cp312-win_amd64.whl", hash = "sha256:ae76c22bde5cbb6bfd211ec343ded2163bba7883c7bc77f6b756a1049436fbb9"},
{file = "black-23.12.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1fa88a0f74e50e4487477bc0bb900c6781dbddfdfa32691e780bf854c3b4a47f"},
{file = "black-23.12.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:a4d6a9668e45ad99d2f8ec70d5c8c04ef4f32f648ef39048d010b0689832ec6d"},
{file = "black-23.12.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b18fb2ae6c4bb63eebe5be6bd869ba2f14fd0259bda7d18a46b764d8fb86298a"},
{file = "black-23.12.1-cp38-cp38-win_amd64.whl", hash = "sha256:c04b6d9d20e9c13f43eee8ea87d44156b8505ca8a3c878773f68b4e4812a421e"},
{file = "black-23.12.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3e1b38b3135fd4c025c28c55ddfc236b05af657828a8a6abe5deec419a0b7055"},
{file = "black-23.12.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4f0031eaa7b921db76decd73636ef3a12c942ed367d8c3841a0739412b260a54"},
{file = "black-23.12.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:97e56155c6b737854e60a9ab1c598ff2533d57e7506d97af5481141671abf3ea"},
{file = "black-23.12.1-cp39-cp39-win_amd64.whl", hash = "sha256:dd15245c8b68fe2b6bd0f32c1556509d11bb33aec9b5d0866dd8e2ed3dba09c2"},
{file = "black-23.12.1-py3-none-any.whl", hash = "sha256:78baad24af0f033958cad29731e27363183e140962595def56423e626f4bee3e"},
{file = "black-23.12.1.tar.gz", hash = "sha256:4ce3ef14ebe8d9509188014d96af1c456a910d5b5cbf434a09fef7e024b3d0d5"},
{file = "black-24.10.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e6668650ea4b685440857138e5fe40cde4d652633b1bdffc62933d0db4ed9812"},
{file = "black-24.10.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1c536fcf674217e87b8cc3657b81809d3c085d7bf3ef262ead700da345bfa6ea"},
{file = "black-24.10.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:649fff99a20bd06c6f727d2a27f401331dc0cc861fb69cde910fe95b01b5928f"},
{file = "black-24.10.0-cp310-cp310-win_amd64.whl", hash = "sha256:fe4d6476887de70546212c99ac9bd803d90b42fc4767f058a0baa895013fbb3e"},
{file = "black-24.10.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5a2221696a8224e335c28816a9d331a6c2ae15a2ee34ec857dcf3e45dbfa99ad"},
{file = "black-24.10.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f9da3333530dbcecc1be13e69c250ed8dfa67f43c4005fb537bb426e19200d50"},
{file = "black-24.10.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4007b1393d902b48b36958a216c20c4482f601569d19ed1df294a496eb366392"},
{file = "black-24.10.0-cp311-cp311-win_amd64.whl", hash = "sha256:394d4ddc64782e51153eadcaaca95144ac4c35e27ef9b0a42e121ae7e57a9175"},
{file = "black-24.10.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:b5e39e0fae001df40f95bd8cc36b9165c5e2ea88900167bddf258bacef9bbdc3"},
{file = "black-24.10.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:d37d422772111794b26757c5b55a3eade028aa3fde43121ab7b673d050949d65"},
{file = "black-24.10.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:14b3502784f09ce2443830e3133dacf2c0110d45191ed470ecb04d0f5f6fcb0f"},
{file = "black-24.10.0-cp312-cp312-win_amd64.whl", hash = "sha256:30d2c30dc5139211dda799758559d1b049f7f14c580c409d6ad925b74a4208a8"},
{file = "black-24.10.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:1cbacacb19e922a1d75ef2b6ccaefcd6e93a2c05ede32f06a21386a04cedb981"},
{file = "black-24.10.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:1f93102e0c5bb3907451063e08b9876dbeac810e7da5a8bfb7aeb5a9ef89066b"},
{file = "black-24.10.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ddacb691cdcdf77b96f549cf9591701d8db36b2f19519373d60d31746068dbf2"},
{file = "black-24.10.0-cp313-cp313-win_amd64.whl", hash = "sha256:680359d932801c76d2e9c9068d05c6b107f2584b2a5b88831c83962eb9984c1b"},
{file = "black-24.10.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:17374989640fbca88b6a448129cd1745c5eb8d9547b464f281b251dd00155ccd"},
{file = "black-24.10.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:63f626344343083322233f175aaf372d326de8436f5928c042639a4afbbf1d3f"},
{file = "black-24.10.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ccfa1d0cb6200857f1923b602f978386a3a2758a65b52e0950299ea014be6800"},
{file = "black-24.10.0-cp39-cp39-win_amd64.whl", hash = "sha256:2cd9c95431d94adc56600710f8813ee27eea544dd118d45896bb734e9d7a0dc7"},
{file = "black-24.10.0-py3-none-any.whl", hash = "sha256:3bb2b7a1f7b685f85b11fed1ef10f8a9148bceb49853e47a294a3dd963c1dd7d"},
{file = "black-24.10.0.tar.gz", hash = "sha256:846ea64c97afe3bc677b761787993be4991810ecc7a4a937816dd6bddedc4875"},
]
[package.dependencies]
@@ -421,7 +421,7 @@ platformdirs = ">=2"
[package.extras]
colorama = ["colorama (>=0.4.3)"]
d = ["aiohttp (>=3.7.4) ; sys_platform != \"win32\" or implementation_name != \"pypy\"", "aiohttp (>=3.7.4,!=3.9.0) ; sys_platform == \"win32\" and implementation_name == \"pypy\""]
d = ["aiohttp (>=3.10)"]
jupyter = ["ipython (>=7.8.0)", "tokenize-rt (>=3.2.0)"]
uvloop = ["uvloop (>=0.15.2)"]
@@ -2984,7 +2984,7 @@ description = "Fast transfer of large files with the Hugging Face Hub."
optional = false
python-versions = ">=3.8"
groups = ["main"]
markers = "(sys_platform == \"linux\" or platform_machine == \"x86_64\" or platform_machine == \"amd64\" or platform_machine == \"AMD64\" or platform_machine == \"arm64\" or platform_machine == \"aarch64\") and (platform_machine == \"x86_64\" or platform_machine == \"aarch64\" or platform_machine == \"amd64\" or platform_machine == \"AMD64\" or platform_machine == \"arm64\")"
markers = "platform_machine == \"x86_64\" or platform_machine == \"amd64\" or platform_machine == \"AMD64\" or platform_machine == \"arm64\" or platform_machine == \"aarch64\""
files = [
{file = "hf_xet-1.2.0-cp313-cp313t-macosx_10_12_x86_64.whl", hash = "sha256:ceeefcd1b7aed4956ae8499e2199607765fbd1c60510752003b6cc0b8413b649"},
{file = "hf_xet-1.2.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:b70218dd548e9840224df5638fdc94bd033552963cfa97f9170829381179c813"},
@@ -8774,4 +8774,4 @@ type = ["pytest-mypy"]
[metadata]
lock-version = "2.1"
python-versions = "^3.12"
content-hash = "bd647e9d41cdc55cb4b3d784746e3e09378180b730e3de3bf3e2df1899c113bd"
content-hash = "9d632f4341ef2e49aa81978032caf9605fd258aff8176263ccb1766b0a1ef4b1"

View File

@@ -111,7 +111,7 @@ pytest = "^8.0"
[tool.poetry.group.dev.dependencies]
# Formatting & linting
black = "^23.12.1"
black = "^24.10.0"
flake8 = "^7.0.0"
isort = "^5.13.1"
pre-commit = "^3.3.3"