Subjects : Multi-instance
| Type | Year | Title | Cited | Download |
|---|---|---|---|---|
| Conference | 2025 | Disk-Based Shared KV Cache Management for Fast Inference in Multi-Instance LLM RAG Systems 이형우 International Conference on Cloud Computing (CLOUD) 2025, pp.1-11 | ||
| Conference | 2024 | ParvaGPU: Efficient Spatial GPU Sharing for Large-Scale DNN Inference in Cloud Environments 이문규 International Conference for High Performance Computing, Networking, Storage, and Analysis (SC) 2024, pp.1-14 | 2 | 원문 |
| Status | Year | Patent Name | Country | Family Pat. | KIPRIS |
|---|---|---|---|---|---|
| No search results. | |||||
| Type | Year | Research Project | Primary Investigator | Download |
|---|---|---|---|---|
| No search results. | ||||