A high-throughput and memory-efficient inference and serving engine for LLMs
- Python
00Обновлено 8 месяцев назад
Operating LLMs in production
llmaillamafine-tuningmistralmlopsllama2llm-inferencellm-servingllmopsmlfalconbentomlllm-opsmodel-inferencemptopen-source-llmopenllmstablelmvicuna- Python
00Обновлено 7 месяцев назад
RayLLM - LLMs on Ray
- Python
00Обновлено 7 месяцев назад
SkyPilot: Run LLMs, AI, and Batch jobs on any cloud. Get maximum savings, highest GPU availability, and managed execution—all with a simple interface.
machine-learningdeep-learningllm-servinggpudata-scienceml-platformdistributed-trainingcloud-computingcloud-managementmulticloudhyperparameter-tuningcost-managementcost-optimizationfinopsjob-queuejob-schedulerllm-trainingml-infrastructurespot-instancestpu- Python
00Обновлено 7 месяцев назад
LLM (Large Language Model) FineTuning
llmlarge-language-modelspytorchllama2gpt-3llm-inferencellm-servingllm-trainingopen-source-llmmistral-7bgpt3-turbollm-finetuning- Jupyter Notebook
00Обновлено 7 месяцев назад