You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
"The Dual PCIe Card, featuring two SAKURA-II accelerators, is ideal for insertion into PCIe backplanes and fits comfortably into a single slot even with the attached heat sink. This card provides significant onboard DRAM, making it optimal for Generative AI and other memory-intensive AI applications like Large Language Models (LLMs).
DRAM Memory: 32GB (4x banks of 8GB LPDDR4)
Form Factor: Half Height, Half Length (HHHL) x16 slot card
Host Interface: PCIe Gen 3.0 x8/x8 (bifurcated)
Peak Performance: 120 TOPS with INT8, 60 TFLOPS with BF16
Card Power: 20W (typical)"
"KP-2 PCIe Card
The Kinara Ara-2 PCIe AI accelerator card enables high performance, power-efficient AI inference for edge server applications including Generative AI workloads such as Stable Diffusion and LLMs.
The KP-2 is powered by four 40 TOPS Ara-2s, rivaling the latest GPU-based inference cards in performance but at the fraction of power and cost.
The KP-2 is ideal for running heavy inference workloads with multiple modern CNNs, transformers, and Generative AI models in a variety of smart city, retail, gaming, and enterprise applications."
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Has anyone bought one of these and tried it out?
i was eyeing the SAKURA-II Dual Eval Card (32GB)
EdgeCortix:
https://www.edgecortix.com/en/hardware
"The Dual PCIe Card, featuring two SAKURA-II accelerators, is ideal for insertion into PCIe backplanes and fits comfortably into a single slot even with the attached heat sink. This card provides significant onboard DRAM, making it optimal for Generative AI and other memory-intensive AI applications like Large Language Models (LLMs).
DRAM Memory: 32GB (4x banks of 8GB LPDDR4)
Form Factor: Half Height, Half Length (HHHL) x16 slot card
Host Interface: PCIe Gen 3.0 x8/x8 (bifurcated)
Peak Performance: 120 TOPS with INT8, 60 TFLOPS with BF16
Card Power: 20W (typical)"
Kinara:
https://kinara.ai/products/kinara-ara-2/
"KP-2 PCIe Card
The Kinara Ara-2 PCIe AI accelerator card enables high performance, power-efficient AI inference for edge server applications including Generative AI workloads such as Stable Diffusion and LLMs.
The KP-2 is powered by four 40 TOPS Ara-2s, rivaling the latest GPU-based inference cards in performance but at the fraction of power and cost.
The KP-2 is ideal for running heavy inference workloads with multiple modern CNNs, transformers, and Generative AI models in a variety of smart city, retail, gaming, and enterprise applications."
Memryx:
https://memryx.com/products/
Available on Amazon
"Each M.2 offers:
4 MX3 chips
Up to 24 TFLOPS (1GHz)
2280 M-Key form factor (22 x 80mm)
42M 8-bit parameters (weights)
x2 or x4 PCIe Gen 3 interface"
Beta Was this translation helpful? Give feedback.
All reactions