2022-07-05 00:02:33 +00:00
|
|
|
from min_dalle import MinDalle
|
2022-06-29 19:53:25 +00:00
|
|
|
import tempfile
|
2022-07-09 10:48:51 +00:00
|
|
|
import torch, torch.backends.cudnn
|
2022-07-05 00:02:33 +00:00
|
|
|
from typing import Iterator
|
2022-06-29 19:53:25 +00:00
|
|
|
from cog import BasePredictor, Path, Input
|
|
|
|
|
2022-07-09 10:48:51 +00:00
|
|
|
torch.backends.cudnn.deterministic = False
|
|
|
|
|
2022-06-29 19:53:25 +00:00
|
|
|
|
2022-07-05 09:47:35 +00:00
|
|
|
class ReplicatePredictor(BasePredictor):
|
2022-06-29 19:53:25 +00:00
|
|
|
def setup(self):
|
2022-07-05 11:07:29 +00:00
|
|
|
self.model = MinDalle(is_mega=True, is_reusable=True)
|
2022-06-29 19:53:25 +00:00
|
|
|
|
|
|
|
def predict(
|
|
|
|
self,
|
2022-07-12 13:17:18 +00:00
|
|
|
text: str = Input(default='Dali painting of WALL·E'),
|
|
|
|
intermediate_outputs: bool = Input(default=True),
|
|
|
|
grid_size: int = Input(ge=1, le=9, default=5),
|
|
|
|
log2_temperature: float = Input(ge=-3, le=3, default=1),
|
|
|
|
log2_top_k: int = Input(ge=0, le=14, default=7),
|
|
|
|
log2_supercondition_factor: int = Input(ge=2, le=6, default=4)
|
2022-07-05 00:02:33 +00:00
|
|
|
) -> Iterator[Path]:
|
2022-07-12 00:58:00 +00:00
|
|
|
try:
|
2022-07-07 16:35:00 +00:00
|
|
|
image_stream = self.model.generate_image_stream(
|
2022-07-11 17:19:06 +00:00
|
|
|
text = text,
|
|
|
|
seed = -1,
|
|
|
|
grid_size = grid_size,
|
|
|
|
log2_mid_count = 3 if intermediate_outputs else 0,
|
2022-07-12 13:17:18 +00:00
|
|
|
temperature = 2 ** log2_temperature,
|
|
|
|
supercondition_factor = 2 ** log2_supercondition_factor,
|
|
|
|
top_k = 2 ** log2_top_k,
|
2022-07-11 17:19:06 +00:00
|
|
|
is_verbose = True
|
2022-07-07 16:35:00 +00:00
|
|
|
)
|
2022-07-04 22:37:07 +00:00
|
|
|
|
2022-07-07 16:35:00 +00:00
|
|
|
iter = 0
|
|
|
|
path = Path(tempfile.mkdtemp())
|
|
|
|
for image in image_stream:
|
|
|
|
iter += 1
|
|
|
|
image_path = path / 'min-dalle-iter-{}.jpg'.format(iter)
|
|
|
|
image.save(str(image_path))
|
|
|
|
yield image_path
|
|
|
|
except:
|
|
|
|
print("An error occured, deleting model")
|
|
|
|
del self.model
|
2022-07-07 21:03:47 +00:00
|
|
|
torch.cuda.empty_cache()
|
2022-07-07 16:35:00 +00:00
|
|
|
self.setup()
|
|
|
|
raise Exception("There was an error, please try again")
|