Skip to content

Commit

Permalink
clean up
Browse files Browse the repository at this point in the history
  • Loading branch information
Kye committed Sep 27, 2023
1 parent ec5f2ad commit 594e720
Show file tree
Hide file tree
Showing 2 changed files with 5 additions and 4 deletions.
8 changes: 5 additions & 3 deletions mmca/main.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,8 @@
import torch
from torch import nn
from einops import rearrange
import torch
import torch.nn.functional as F
from einops import rearrange
from torch import nn

# from zeta.nn import FlashAttention

class MultiModalCausalAttention(nn.Module):
Expand Down Expand Up @@ -103,6 +104,7 @@ def __init__(
embed_dim=dim,
num_heads=heads
)

self.cross_attn = nn.MultiheadAttention(
embed_dim=dim,
num_heads=heads
Expand Down
1 change: 0 additions & 1 deletion simple_example.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,6 @@
import torch
from mmca.main import SimpleMMCA


# Define the dimensions
dim = 512
head = 8
Expand Down

0 comments on commit 594e720

Please sign in to comment.