EmbeddedLLM Platform Team
Oct 27, 2023
High throughput LLM inference with vLLM and AMD: Achieving LLM inference parity with Nvidia
EmbeddedLLM has ported vLLM to ROCm 5.6, and we are excited to report that LLM inference has achieved parity with Nvidia A100 using AMD MI210.
BY JIAQI LIM
Oct 20, 2023
Real-Time Flood Detection: Achieving Supply Chain Resilience through Large Language Model and Image Analysis 🚗