[Electron] Add fp32 and fp64 to unet inference precision options (#1678)

This commit is contained in:
Chenlei Hu
2024-11-25 08:20:49 -08:00
committed by GitHub
parent 43c23e526c
commit 3708afaf21
2 changed files with 3 additions and 0 deletions

View File

@@ -163,6 +163,8 @@ export const SERVER_CONFIG_ITEMS: ServerConfig<any>[] = [
type: 'combo',
options: [
FloatingPointPrecision.AUTO,
FloatingPointPrecision.FP64,
FloatingPointPrecision.FP32,
FloatingPointPrecision.FP16,
FloatingPointPrecision.BF16,
FloatingPointPrecision.FP8E4M3FN,

View File

@@ -40,6 +40,7 @@ export enum CudaMalloc {
export enum FloatingPointPrecision {
AUTO = 'auto',
FP64 = 'fp64',
FP32 = 'fp32',
FP16 = 'fp16',
BF16 = 'bf16',