Jeli itko probao koristiti RAMDisk za lokalni AI deployment tipa DeepSeek ili nes slicno? 
 
Postavio sam upit onom drugom statistickom pogadjalu (DeepSeek, jel...) za sto bi se dalo koristiti RAMDisk kod lokalne instalacije i ovo mi je napisalo... 
 
 
What DeepSeek Can Use the RAMDisk For 
 
Model Cache: 
- Store frequently accessed model weights (if using llama.cpp or similar). 
 
Temporary File Storage: 
- Store intermediate files (logs, tokenized inputs). 
 
Swap for Small-Batch Inference: 
- If your system has limited RAM, a RAMDisk can help avoid SSD thrashing. 
 
Faster Disk-Based KV Caching: 
- Some inference engines allow offloading cache to disk - use RAMDisk for speed. 
 
 
 
Best for: 
- Small models (e.g., deepseek-coder-1.3b in 4-bit) that fit entirely in RAMDisk. 
- Frequent disk I/O tasks (like caching prompts).  
		
	
		
		
		
		
		
		
			
				__________________ 
				 
GMKTec K8Plus 
AMD Ryzen 7 PRO 8845HS, 3.80-5.10GHz 8-core 
64GB DDR5-5200 (2x32GB), quad-channel 
NVMe1 - 512GB Micron 3400, PCI-E 4.0 x4 
NVMe2 - 512GB Samsung PM981, PCI-E 3.0 x4 
AMD Radeon 780M integrirana grafika 
 
 
2x EIZO EV2336WFS3-BK 
23" 1920x1080 IPS 
 
 
Dell Latitude 7390 
intel i7 8650U, 1.90-4.20GHz, 4-core 
32GB DDR4-2666 
512GB Samsung PM981a, PCI-E 3.0 x4 
intel UHD 620 
13.3" FullHD IPS Touchscreen 
 
 
			 
		
		
		
		
		
		
	
	 |