Merge branch 'v2.3' into feat/lora-support-2.3

This commit is contained in:
Lincoln Stein
2023-04-01 10:55:22 -04:00
committed by GitHub
4 changed files with 12 additions and 11 deletions

View File

@@ -77,7 +77,7 @@ machine. To test, open up a terminal window and issue the following
command:
```
rocm-smi
rocminfo
```
If you get a table labeled "ROCm System Management Interface" the
@@ -95,16 +95,18 @@ recent version of Ubuntu, 22.04. However, this [community-contributed
recipe](https://novaspirit.github.io/amdgpu-rocm-ubu22/) is reported
to work well.
After installation, please run `rocm-smi` a second time to confirm
After installation, please run `rocminfo` a second time to confirm
that the driver is present and the GPU is recognized. You may need to
do a reboot in order to load the driver. If you continue to have
errors you may need to add your username to the system `render`
group. This can be done with the following command:
do a reboot in order to load the driver. In addition, if you see
errors relating to your username not being a member of the `render`
group, you may fix this by adding yourself to this group with the command:
```
sudo usermod -a -G render my-user-name
sudo usermod -a -G render myUserName
```
(Thanks to @EgoringKosmos for the usermod recipe.)
### Linux Install with a ROCm-docker Container
If you are comfortable with the Docker containerization system, then

View File

@@ -242,8 +242,8 @@ class InvokeAiInstance:
from plumbum import FG, local
# Note that we're installing pinned versions of torch and
# torchvision here, which may not correspond to what is
# in pyproject.toml. This is a hack to prevent torch 2.0 from
# torchvision here, which *should* correspond to what is
# in pyproject.toml. This is to prevent torch 2.0 from
# being installed and immediately uninstalled and replaced with 1.13
pip = local[self.pip]
@@ -252,7 +252,7 @@ class InvokeAiInstance:
"install",
"--require-virtualenv",
"torch~=1.13.1",
"torchvision>=0.14.1",
"torchvision~=0.14.1",
"--force-reinstall",
"--find-links" if find_links is not None else None,
find_links,

View File

@@ -78,7 +78,7 @@ def main():
elif choice=='4':
branch = Prompt.ask('Enter an InvokeAI branch name')
print(f':crossed_fingers: Upgrading to [yellow]{tag}[/yellow]')
print(f':crossed_fingers: Upgrading to [yellow]{tag if tag else release}[/yellow]')
if release:
cmd = f'pip install {INVOKE_AI_SRC}/{release}.zip --use-pep517 --upgrade'
elif tag:

View File

@@ -464,7 +464,6 @@ class FrozenCLIPEmbedder(AbstractEncoder):
return self(text, **kwargs)
def set_textual_inversion_manager(self, manager): #TextualInversionManager):
# TODO all of the weighting and expanding stuff needs be moved out of this class
self.textual_inversion_manager = manager
@property