{"labels":{"en":"Deploying LLMs on CPU-only Environments with llama.cpp Library Set: MedLocalGPT Project Case"},"descriptions":{"en":"scientific paper published in CEUR-WS Volume 4164"},"claims":{"P31":"Q13442814","P1433":null,"P1476":{"text":"Deploying LLMs on CPU-only Environments with llama.cpp Library Set: MedLocalGPT Project Case","language":"en"},"P407":"Q1860","P953":"https://ceur-ws.org/Vol-4164/paper11.pdf","P50":[],"P2093":[{"value":"Kyrylo S. Malakhov","qualifiers":{"P1545":"1"}}]}}