mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	llama : increase scratch buffer size for 65B (ref #1152)
Temporary solution
This commit is contained in:
		| @@ -54,7 +54,7 @@ static const std::map<e_model, size_t> & MEM_REQ_SCRATCH0() | |||||||
|         { MODEL_7B,    512ull * MB }, |         { MODEL_7B,    512ull * MB }, | ||||||
|         { MODEL_13B,   512ull * MB }, |         { MODEL_13B,   512ull * MB }, | ||||||
|         { MODEL_30B,   512ull * MB }, |         { MODEL_30B,   512ull * MB }, | ||||||
|         { MODEL_65B,   512ull * MB }, |         { MODEL_65B,  1024ull * MB }, | ||||||
|     }; |     }; | ||||||
|     return _MEM_REQ_SCRATCH0; |     return _MEM_REQ_SCRATCH0; | ||||||
| } | } | ||||||
| @@ -65,7 +65,7 @@ static const std::map<e_model, size_t> & MEM_REQ_SCRATCH1() | |||||||
|         { MODEL_7B,    512ull * MB }, |         { MODEL_7B,    512ull * MB }, | ||||||
|         { MODEL_13B,   512ull * MB }, |         { MODEL_13B,   512ull * MB }, | ||||||
|         { MODEL_30B,   512ull * MB }, |         { MODEL_30B,   512ull * MB }, | ||||||
|         { MODEL_65B,   512ull * MB }, |         { MODEL_65B,  1024ull * MB }, | ||||||
|     }; |     }; | ||||||
|     return _MEM_REQ_SCRATCH1; |     return _MEM_REQ_SCRATCH1; | ||||||
| } | } | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov