You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
61 lines
2.1 KiB
61 lines
2.1 KiB
from min_dalle import MinDalle |
|
import tempfile |
|
import torch, torch.backends.cudnn |
|
from typing import Iterator |
|
from cog import BasePredictor, Path, Input |
|
|
|
torch.backends.cudnn.deterministic = False |
|
|
|
|
|
class ReplicatePredictor(BasePredictor): |
|
def setup(self): |
|
self.model = MinDalle(is_mega=True, is_reusable=True) |
|
|
|
def predict( |
|
self, |
|
text: str = Input( |
|
description='For long prompts, only the first 64 tokens will be used to generate the image.', |
|
default='Dali painting of WALL·E' |
|
), |
|
intermediate_outputs: bool = Input( |
|
description='Whether to show intermediate outputs while running. This adds less than a second to the run time.', |
|
default=True |
|
), |
|
grid_size: int = Input( |
|
description='Size of the image grid. 4x4 takes about 15 seconds, 8x8 takes about 35 seconds', |
|
ge=1, |
|
le=8, |
|
default=4 |
|
), |
|
log2_supercondition_factor: int = Input( |
|
description='Higher values result in better agreement with the text but a narrower variety of generated images', |
|
ge=1, |
|
le=6, |
|
default=4 |
|
), |
|
) -> Iterator[Path]: |
|
try: |
|
seed = -1 |
|
log2_mid_count = 3 if intermediate_outputs else 0 |
|
image_stream = self.model.generate_image_stream( |
|
text, |
|
seed, |
|
grid_size=grid_size, |
|
log2_mid_count=log2_mid_count, |
|
log2_supercondition_factor=log2_supercondition_factor, |
|
is_verbose=True |
|
) |
|
|
|
iter = 0 |
|
path = Path(tempfile.mkdtemp()) |
|
for image in image_stream: |
|
iter += 1 |
|
image_path = path / 'min-dalle-iter-{}.jpg'.format(iter) |
|
image.save(str(image_path)) |
|
yield image_path |
|
except: |
|
print("An error occured, deleting model") |
|
del self.model |
|
torch.cuda.empty_cache() |
|
self.setup() |
|
raise Exception("There was an error, please try again") |