Joanne Lei

Joanne Lei is a Solution Architect whose work focuses on designing, building, and deploying Cloud native application and infrastructure, with specialization in optimizing AIML workloads on GPU and deploying them in production. She earned a BSECE degree from UT Austin. 

Recent Blogs

Serving LLM using HuggingFace and Kubernetes on OCI - Part II

As a follow-up to a previous post, this dives deeper into building Docker images for model-downloader and the web inference client, enabling an interactive demo for your fine-tuned model.

Serving LLMs using HuggingFace and Kubernetes on OCI

In this blog post, I will demonstrate how to deploy fine-tuned LLM inference containers on OKE, an OCI managed Kubernetes service that simplifies deployments and operations at scale for enterprises.

Serve ML models at scale with NVIDIA Triton Inference Server on OKE

In this blog, you will find out how to deploy ML models at scale to deliver high performing and cost-effective inference service on OCI.

  1. View more

Receive the latest blog updates