Co-authored-by: graemeniedermayer graemeniedermayer@users.noreply.github.com
This commit is contained in:
DenOfEquity
2025-02-27 17:54:44 +00:00
committed by GitHub
parent 8dd92501e6
commit f23bc80d2f
6 changed files with 1184 additions and 22 deletions

View File

@@ -2,11 +2,13 @@ import torch
import math
from backend.attention import attention_pytorch as attention_function
from transformers.activations import NewGELUActivation
activations = {
"gelu_pytorch_tanh": lambda a: torch.nn.functional.gelu(a, approximate="tanh"),
"relu": torch.nn.functional.relu,
"gelu_new": lambda a: NewGELUActivation()(a),
}