Update iris/src/tokenizer.py
Browse files- iris/src/tokenizer.py +2 -2
iris/src/tokenizer.py
CHANGED
|
@@ -8,8 +8,8 @@ from einops import rearrange
|
|
| 8 |
import torch
|
| 9 |
import torch.nn as nn
|
| 10 |
|
| 11 |
-
from models.lpips import LPIPS
|
| 12 |
-
from models.nets import Encoder, Decoder
|
| 13 |
|
| 14 |
class Tokenizer(nn.Module):
|
| 15 |
def __init__(self, vocab_size: int, embed_dim: int, encoder: Encoder, decoder: Decoder, with_lpips: bool = True) -> None:
|
|
|
|
| 8 |
import torch
|
| 9 |
import torch.nn as nn
|
| 10 |
|
| 11 |
+
from .models.lpips import LPIPS
|
| 12 |
+
from .models.nets import Encoder, Decoder
|
| 13 |
|
| 14 |
class Tokenizer(nn.Module):
|
| 15 |
def __init__(self, vocab_size: int, embed_dim: int, encoder: Encoder, decoder: Decoder, with_lpips: bool = True) -> None:
|