Texonom
Texonom
/
Engineering
Engineering
/Data Engineering/Artificial Intelligence/AI Development/AI Optimization/Inference Optimization/Unstructured Sparsity/
Flash-LLM
Search

Flash-LLM

Creator
Creator
Seonglae Cho
Created
Created
2023 Sep 20 16:27
Editor
Editor
Seonglae Cho
Edited
Edited
2023 Sep 20 16:28
Refs
Refs
Tensor Core
flash-llm
AlibabaResearch • Updated 2024 Jul 12 13:56
 
 
 
 
 
 
 
Papers with Code - Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
Implemented in one code library.
Papers with Code - Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
https://paperswithcode.com/paper/flash-llm-enabling-cost-effective-and-highly
Papers with Code - Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
 
 

 
 
 
 
 
 
 
 
 

Recommendations

Texonom
Texonom
/
Engineering
Engineering
/Data Engineering/Artificial Intelligence/AI Development/AI Optimization/Inference Optimization/Unstructured Sparsity/
Flash-LLM
Copyright Seonglae Cho