Skip to content

Instantly share code, notes, and snippets.

View TheaperDeng's full-sized avatar
🍦
Chilling

Junwei Deng TheaperDeng

🍦
Chilling
View GitHub Profile
@TheaperDeng
TheaperDeng / dattri-notebook-3.ipynb
Last active October 28, 2024 02:56
dattri-notebook-3.ipynb
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
@TheaperDeng
TheaperDeng / dattri-notebook-2.ipynb
Created October 9, 2024 20:43
Dattri-notebook-2.ipynb
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
@TheaperDeng
TheaperDeng / dattri-notebook-1.ipynb
Created October 9, 2024 20:29
Dattri-notebook-1.ipynb
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.

ggml-gpu-hands-on

This gist works as a hands-on note for running llama.cpp on various GPU. It may out-of-date due to the proj update.

This is only a personal record so readers may not have an out-of-box hands-on experience

Complete configs

Record the verified configs. The project is still developing very fast, so the granularity for the record is specified to the commit id. | Imple. | Device | OS | llama.cpp version | 3rd party version | Step | | ------ | --------------------------------- | ------------ | ------------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------

Summary

General Motivation

background: In Ray AI Runtime (AIR), Trainer is used to scale deep learning/machine learning model's training and tuning process. There is a TorchTrainer for (data parallel / local) pytorch (training / tunning) by providing customized train_loop_per_worker.

motivation1: pytorch-lightning is getting popular among AI researchers and machine learning engineers. The LightningModule in pytorch-lightning contains the customized training loop. It could be user-friendly and more straight-forward if users can only input a LightningModule to a Ray AIR Trainer's initialization process and generate an instance of PytorchLightningTrainer for distributed training / hyperparameter tuning.

motivation2: CPU training is a valid choice for AI models' training, while currently CPU training acceleration is not emphasized. bigdl-nano is a python package to transparently accelerate Pytorch and Tensorflow applicat