jetson-examples 0.1.8__py3-none-any.whl → 0.1.9__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {jetson_examples-0.1.8.dist-info → jetson_examples-0.1.9.dist-info}/LICENSE +21 -21
- {jetson_examples-0.1.8.dist-info → jetson_examples-0.1.9.dist-info}/METADATA +1 -1
- jetson_examples-0.1.9.dist-info/RECORD +109 -0
- reComputer/__init__.py +1 -1
- reComputer/main.py +60 -60
- reComputer/scripts/MoveNet-Lightning/clean.sh +8 -8
- reComputer/scripts/MoveNet-Lightning/getVersion.sh +59 -59
- reComputer/scripts/MoveNet-Lightning/init.sh +6 -6
- reComputer/scripts/MoveNet-Lightning/readme.md +30 -30
- reComputer/scripts/MoveNet-Lightning/run.sh +19 -19
- reComputer/scripts/MoveNet-Thunder/clean.sh +7 -7
- reComputer/scripts/MoveNet-Thunder/getVersion.sh +59 -59
- reComputer/scripts/MoveNet-Thunder/init.sh +6 -6
- reComputer/scripts/MoveNet-Thunder/readme.md +31 -31
- reComputer/scripts/MoveNet-Thunder/run.sh +18 -18
- reComputer/scripts/MoveNetJS/clean.sh +4 -4
- reComputer/scripts/MoveNetJS/readme.md +56 -56
- reComputer/scripts/MoveNetJS/run.sh +13 -13
- reComputer/scripts/Sheared-LLaMA-2.7B-ShareGPT/init.sh +16 -16
- reComputer/scripts/Sheared-LLaMA-2.7B-ShareGPT/run.sh +8 -8
- reComputer/scripts/audiocraft/README.md +35 -35
- reComputer/scripts/audiocraft/clean.sh +5 -5
- reComputer/scripts/audiocraft/init.sh +16 -16
- reComputer/scripts/audiocraft/run.sh +7 -7
- reComputer/scripts/check.sh +4 -4
- reComputer/scripts/clean.sh +33 -33
- reComputer/scripts/comfyui/LICENSE +21 -21
- reComputer/scripts/comfyui/README.md +127 -127
- reComputer/scripts/comfyui/clean.sh +9 -7
- reComputer/scripts/comfyui/config.yaml +30 -29
- reComputer/scripts/comfyui/init.sh +9 -163
- reComputer/scripts/comfyui/run.sh +30 -30
- reComputer/scripts/depth-anything/Dockerfile +5 -5
- reComputer/scripts/depth-anything/LICENSE +21 -21
- reComputer/scripts/depth-anything/README.md +135 -135
- reComputer/scripts/depth-anything/clean.sh +7 -7
- reComputer/scripts/depth-anything/config.yaml +31 -31
- reComputer/scripts/depth-anything/init.sh +164 -164
- reComputer/scripts/depth-anything/run.sh +22 -22
- reComputer/scripts/depth-anything-v2/Dockerfile +5 -5
- reComputer/scripts/depth-anything-v2/LICENSE +21 -21
- reComputer/scripts/depth-anything-v2/README.md +135 -135
- reComputer/scripts/depth-anything-v2/clean.sh +7 -7
- reComputer/scripts/depth-anything-v2/config.yaml +31 -31
- reComputer/scripts/depth-anything-v2/init.sh +164 -164
- reComputer/scripts/depth-anything-v2/run.sh +22 -22
- reComputer/scripts/live-llava/init.sh +16 -16
- reComputer/scripts/live-llava/run.sh +278 -278
- reComputer/scripts/llama-factory/README.md +68 -68
- reComputer/scripts/llama-factory/clean.sh +4 -4
- reComputer/scripts/llama-factory/init.sh +52 -52
- reComputer/scripts/llama-factory/run.sh +10 -10
- reComputer/scripts/llama3/clean.sh +22 -22
- reComputer/scripts/llama3/config.yaml +31 -0
- reComputer/scripts/llama3/init.sh +19 -16
- reComputer/scripts/llama3/run.sh +13 -13
- reComputer/scripts/llava/clean.sh +3 -3
- reComputer/scripts/llava/init.sh +16 -16
- reComputer/scripts/llava/run.sh +9 -9
- reComputer/scripts/llava-v1.5-7b/init.sh +16 -16
- reComputer/scripts/llava-v1.5-7b/run.sh +9 -9
- reComputer/scripts/llava-v1.6-vicuna-7b/init.sh +16 -16
- reComputer/scripts/llava-v1.6-vicuna-7b/run.sh +10 -10
- reComputer/scripts/nanodb/init.sh +16 -16
- reComputer/scripts/nanodb/readme.md +10 -10
- reComputer/scripts/nanodb/run.sh +90 -90
- reComputer/scripts/nanoowl/init.sh +16 -16
- reComputer/scripts/nanoowl/run.sh +7 -7
- reComputer/scripts/ollama/clean.sh +22 -22
- reComputer/scripts/ollama/config.yaml +31 -0
- reComputer/scripts/ollama/init.sh +19 -16
- reComputer/scripts/ollama/run.sh +10 -10
- reComputer/scripts/parler-tts/clean.sh +7 -7
- reComputer/scripts/parler-tts/getVersion.sh +59 -59
- reComputer/scripts/parler-tts/init.sh +8 -8
- reComputer/scripts/parler-tts/readme.md +63 -63
- reComputer/scripts/parler-tts/run.sh +17 -17
- reComputer/scripts/run.sh +48 -48
- reComputer/scripts/stable-diffusion-webui/init.sh +16 -16
- reComputer/scripts/stable-diffusion-webui/run.sh +6 -6
- reComputer/scripts/text-generation-webui/init.sh +16 -16
- reComputer/scripts/text-generation-webui/run.sh +11 -11
- reComputer/scripts/ultralytics-yolo/LICENSE +21 -21
- reComputer/scripts/ultralytics-yolo/README.md +124 -124
- reComputer/scripts/ultralytics-yolo/clean.sh +6 -6
- reComputer/scripts/ultralytics-yolo/config.yaml +31 -31
- reComputer/scripts/ultralytics-yolo/init.sh +4 -4
- reComputer/scripts/ultralytics-yolo/run.sh +26 -26
- reComputer/scripts/update.sh +26 -26
- reComputer/scripts/utils.sh +168 -166
- reComputer/scripts/whisper/init.sh +16 -16
- reComputer/scripts/whisper/run.sh +7 -7
- reComputer/scripts/yolov10/Dockerfile +13 -13
- reComputer/scripts/yolov10/README.md +71 -71
- reComputer/scripts/yolov10/clean.sh +4 -4
- reComputer/scripts/yolov10/config.yaml +31 -31
- reComputer/scripts/yolov10/init.sh +20 -20
- reComputer/scripts/yolov10/run.sh +7 -7
- reComputer/scripts/yolov8-rail-inspection/config.yaml +31 -31
- reComputer/scripts/yolov8-rail-inspection/init.sh +5 -5
- reComputer/scripts/yolov8-rail-inspection/readme.md +35 -35
- reComputer/scripts/yolov8-rail-inspection/run.sh +21 -21
- jetson_examples-0.1.8.dist-info/RECORD +0 -107
- {jetson_examples-0.1.8.dist-info → jetson_examples-0.1.9.dist-info}/WHEEL +0 -0
- {jetson_examples-0.1.8.dist-info → jetson_examples-0.1.9.dist-info}/entry_points.txt +0 -0
- {jetson_examples-0.1.8.dist-info → jetson_examples-0.1.9.dist-info}/top_level.txt +0 -0
@@ -1,68 +1,68 @@
|
|
1
|
-
# Finetune LLM by Llama-Factory on Jetson
|
2
|
-
|
3
|
-
|
4
|
-
## Hello
|
5
|
-
Now you can tailor a custom private local LLM to meet your requirements.
|
6
|
-
|
7
|
-
💡 Here's an example of quickly deploying [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) on Jetson device.
|
8
|
-
|
9
|
-
🔥 Hightlights:
|
10
|
-
- **Llama-Factory** is an efficient tool to unify efficient Fine-Tuning of 100+ LLMs. 🚀🔍
|
11
|
-
- **Jetson-examples** is a toolkit designed to deploy containerized applications on NVIDIA Jetson devices. ✨
|
12
|
-
- **Jetson** is powerful AI hardware platform for edge computing.💻
|
13
|
-
|
14
|
-
🛠️ Follow the tutorial below to quickly experience the performance of Llama-Factory on edge computing devices.
|
15
|
-
|
16
|
-
<div align="center">
|
17
|
-
<img alt="training" width="1200px" src="./assets/training.gif">
|
18
|
-
</div>
|
19
|
-
|
20
|
-
## Get a Jetson Orin Device 🛒
|
21
|
-
| Device Model | Description | Link |
|
22
|
-
|--------------|-------------|------|
|
23
|
-
| reComputer J4012, powered by Orin NX 16GB, 100 TOPS | Embedded computer powered by Orin NX | [Buy Here](https://www.seeedstudio.com/reComputer-J4012-p-5586.html) |
|
24
|
-
| NVIDIA® Jetson AGX Orin™ 64GB Developer Kit | Smallest and most powerful AI edge computer | [Buy Here](https://www.seeedstudio.com/NVIDIArJetson-AGX-Orintm-64GB-Developer-Kit-p-5641.html) |
|
25
|
-
|
26
|
-
## Getting Started
|
27
|
-
|
28
|
-
- install **jetson-examples** by pip:
|
29
|
-
```sh
|
30
|
-
pip3 install jetson-examples
|
31
|
-
```
|
32
|
-
- restart reComputer
|
33
|
-
```sh
|
34
|
-
sudo restart
|
35
|
-
```
|
36
|
-
- run Llama-Factory webui on jetson in one line:
|
37
|
-
```sh
|
38
|
-
reComputer run llama-factory
|
39
|
-
```
|
40
|
-
- Please visit http://127.0.0.1:7860
|
41
|
-
|
42
|
-
<div align="center">
|
43
|
-
<img alt="yolov10" width="1200px" src="./assets/webui.png">
|
44
|
-
</div>
|
45
|
-
|
46
|
-
|
47
|
-
|
48
|
-
## Run Training Script
|
49
|
-
|
50
|
-
> **Note:** Some models and datasets require confirmation before using them, so we recommend logging in with your Hugging Face account by:
|
51
|
-
> `sudo docker exec -it llama-factory huggingface-cli login`
|
52
|
-
|
53
|
-
There are a lot of parameters to choose from webui, you can refer to here for more information.
|
54
|
-
|
55
|
-
For demonstration purposes, set `Model name: Phi-1.5-1.3B`, `Dataset: alpaca_zh`, leave the other parameters unchanged, and then click the `Start` button
|
56
|
-
|
57
|
-
<div align="center">
|
58
|
-
<img alt="yolov10" width="1200px" src="./assets/llama-factory-Jetson.png">
|
59
|
-
</div>
|
60
|
-
|
61
|
-
|
62
|
-
## Build Docker Image
|
63
|
-
We highly recommend that you use `jetson-containers` to compile the docker container, as you can see [here](https://github.com/dusty-nv/jetson-containers/pull/566).
|
64
|
-
|
65
|
-
## Reference
|
66
|
-
- https://github.com/hiyouga/LLaMA-Factory
|
67
|
-
- https://github.com/dusty-nv/jetson-containers
|
68
|
-
|
1
|
+
# Finetune LLM by Llama-Factory on Jetson
|
2
|
+
|
3
|
+
|
4
|
+
## Hello
|
5
|
+
Now you can tailor a custom private local LLM to meet your requirements.
|
6
|
+
|
7
|
+
💡 Here's an example of quickly deploying [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) on Jetson device.
|
8
|
+
|
9
|
+
🔥 Hightlights:
|
10
|
+
- **Llama-Factory** is an efficient tool to unify efficient Fine-Tuning of 100+ LLMs. 🚀🔍
|
11
|
+
- **Jetson-examples** is a toolkit designed to deploy containerized applications on NVIDIA Jetson devices. ✨
|
12
|
+
- **Jetson** is powerful AI hardware platform for edge computing.💻
|
13
|
+
|
14
|
+
🛠️ Follow the tutorial below to quickly experience the performance of Llama-Factory on edge computing devices.
|
15
|
+
|
16
|
+
<div align="center">
|
17
|
+
<img alt="training" width="1200px" src="./assets/training.gif">
|
18
|
+
</div>
|
19
|
+
|
20
|
+
## Get a Jetson Orin Device 🛒
|
21
|
+
| Device Model | Description | Link |
|
22
|
+
|--------------|-------------|------|
|
23
|
+
| reComputer J4012, powered by Orin NX 16GB, 100 TOPS | Embedded computer powered by Orin NX | [Buy Here](https://www.seeedstudio.com/reComputer-J4012-p-5586.html) |
|
24
|
+
| NVIDIA® Jetson AGX Orin™ 64GB Developer Kit | Smallest and most powerful AI edge computer | [Buy Here](https://www.seeedstudio.com/NVIDIArJetson-AGX-Orintm-64GB-Developer-Kit-p-5641.html) |
|
25
|
+
|
26
|
+
## Getting Started
|
27
|
+
|
28
|
+
- install **jetson-examples** by pip:
|
29
|
+
```sh
|
30
|
+
pip3 install jetson-examples
|
31
|
+
```
|
32
|
+
- restart reComputer
|
33
|
+
```sh
|
34
|
+
sudo restart
|
35
|
+
```
|
36
|
+
- run Llama-Factory webui on jetson in one line:
|
37
|
+
```sh
|
38
|
+
reComputer run llama-factory
|
39
|
+
```
|
40
|
+
- Please visit http://127.0.0.1:7860
|
41
|
+
|
42
|
+
<div align="center">
|
43
|
+
<img alt="yolov10" width="1200px" src="./assets/webui.png">
|
44
|
+
</div>
|
45
|
+
|
46
|
+
|
47
|
+
|
48
|
+
## Run Training Script
|
49
|
+
|
50
|
+
> **Note:** Some models and datasets require confirmation before using them, so we recommend logging in with your Hugging Face account by:
|
51
|
+
> `sudo docker exec -it llama-factory huggingface-cli login`
|
52
|
+
|
53
|
+
There are a lot of parameters to choose from webui, you can refer to here for more information.
|
54
|
+
|
55
|
+
For demonstration purposes, set `Model name: Phi-1.5-1.3B`, `Dataset: alpaca_zh`, leave the other parameters unchanged, and then click the `Start` button
|
56
|
+
|
57
|
+
<div align="center">
|
58
|
+
<img alt="yolov10" width="1200px" src="./assets/llama-factory-Jetson.png">
|
59
|
+
</div>
|
60
|
+
|
61
|
+
|
62
|
+
## Build Docker Image
|
63
|
+
We highly recommend that you use `jetson-containers` to compile the docker container, as you can see [here](https://github.com/dusty-nv/jetson-containers/pull/566).
|
64
|
+
|
65
|
+
## Reference
|
66
|
+
- https://github.com/hiyouga/LLaMA-Factory
|
67
|
+
- https://github.com/dusty-nv/jetson-containers
|
68
|
+
|
@@ -1,4 +1,4 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
sudo docker rmi youjiang9977/llama-factory:r35.4.1
|
4
|
-
sudo rm -rf /home/$USER/reComputer/jetson-containers/LLaMA-Factory/*
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
sudo docker rmi youjiang9977/llama-factory:r35.4.1
|
4
|
+
sudo rm -rf /home/$USER/reComputer/jetson-containers/LLaMA-Factory/*
|
@@ -1,52 +1,52 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
|
4
|
-
# check L4T_VERSION
|
5
|
-
# reference https://github.com/dusty-nv/jetson-containers/blob/master/jetson_containers/l4t_version.sh
|
6
|
-
ALLOWED_L4T_VERSIONS=("35.3.1" "35.4.1" "35.5.0")
|
7
|
-
|
8
|
-
ARCH=$(uname -i)
|
9
|
-
if [ $ARCH = "aarch64" ]; then
|
10
|
-
L4T_VERSION_STRING=$(head -n 1 /etc/nv_tegra_release)
|
11
|
-
|
12
|
-
if [ -z "$L4T_VERSION_STRING" ]; then
|
13
|
-
L4T_VERSION_STRING=$(dpkg-query --showformat='${Version}' --show nvidia-l4t-core)
|
14
|
-
L4T_VERSION_ARRAY=(${L4T_VERSION_STRING//./ })
|
15
|
-
L4T_RELEASE=${L4T_VERSION_ARRAY[0]}
|
16
|
-
L4T_REVISION=${L4T_VERSION_ARRAY[1]}
|
17
|
-
else
|
18
|
-
L4T_RELEASE=$(echo $L4T_VERSION_STRING | cut -f 2 -d ' ' | grep -Po '(?<=R)[^;]+')
|
19
|
-
L4T_REVISION=$(echo $L4T_VERSION_STRING | cut -f 2 -d ',' | grep -Po '(?<=REVISION: )[^;]+')
|
20
|
-
fi
|
21
|
-
|
22
|
-
L4T_REVISION_MAJOR=${L4T_REVISION:0:1}
|
23
|
-
L4T_REVISION_MINOR=${L4T_REVISION:2:1}
|
24
|
-
|
25
|
-
L4T_VERSION="$L4T_RELEASE.$L4T_REVISION"
|
26
|
-
echo $L4T_VERSION
|
27
|
-
elif [ $ARCH != "x86_64" ]; then
|
28
|
-
echo "unsupported architecture: $ARCH"
|
29
|
-
exit 1
|
30
|
-
fi
|
31
|
-
|
32
|
-
echo "L4T_VERSION: $L4T_VERSION"
|
33
|
-
if [[ ! " ${ALLOWED_L4T_VERSIONS[@]} " =~ " ${L4T_VERSION} " ]]; then
|
34
|
-
echo "L4T_VERSION is not in the allowed versions list. Exiting."
|
35
|
-
exit 1
|
36
|
-
fi
|
37
|
-
|
38
|
-
|
39
|
-
BASE_PATH=/home/$USER/reComputer
|
40
|
-
mkdir -p $BASE_PATH/
|
41
|
-
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
42
|
-
BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
|
43
|
-
if [ -d $JETSON_REPO_PATH ]; then
|
44
|
-
echo "jetson-ai-lab existed."
|
45
|
-
else
|
46
|
-
echo "jetson-ai-lab does not installed. start init..."
|
47
|
-
cd $BASE_PATH/
|
48
|
-
git clone --depth=1 $BASE_JETSON_LAB_GIT
|
49
|
-
cd $JETSON_REPO_PATH
|
50
|
-
bash install.sh
|
51
|
-
fi
|
52
|
-
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
|
4
|
+
# check L4T_VERSION
|
5
|
+
# reference https://github.com/dusty-nv/jetson-containers/blob/master/jetson_containers/l4t_version.sh
|
6
|
+
ALLOWED_L4T_VERSIONS=("35.3.1" "35.4.1" "35.5.0")
|
7
|
+
|
8
|
+
ARCH=$(uname -i)
|
9
|
+
if [ $ARCH = "aarch64" ]; then
|
10
|
+
L4T_VERSION_STRING=$(head -n 1 /etc/nv_tegra_release)
|
11
|
+
|
12
|
+
if [ -z "$L4T_VERSION_STRING" ]; then
|
13
|
+
L4T_VERSION_STRING=$(dpkg-query --showformat='${Version}' --show nvidia-l4t-core)
|
14
|
+
L4T_VERSION_ARRAY=(${L4T_VERSION_STRING//./ })
|
15
|
+
L4T_RELEASE=${L4T_VERSION_ARRAY[0]}
|
16
|
+
L4T_REVISION=${L4T_VERSION_ARRAY[1]}
|
17
|
+
else
|
18
|
+
L4T_RELEASE=$(echo $L4T_VERSION_STRING | cut -f 2 -d ' ' | grep -Po '(?<=R)[^;]+')
|
19
|
+
L4T_REVISION=$(echo $L4T_VERSION_STRING | cut -f 2 -d ',' | grep -Po '(?<=REVISION: )[^;]+')
|
20
|
+
fi
|
21
|
+
|
22
|
+
L4T_REVISION_MAJOR=${L4T_REVISION:0:1}
|
23
|
+
L4T_REVISION_MINOR=${L4T_REVISION:2:1}
|
24
|
+
|
25
|
+
L4T_VERSION="$L4T_RELEASE.$L4T_REVISION"
|
26
|
+
echo $L4T_VERSION
|
27
|
+
elif [ $ARCH != "x86_64" ]; then
|
28
|
+
echo "unsupported architecture: $ARCH"
|
29
|
+
exit 1
|
30
|
+
fi
|
31
|
+
|
32
|
+
echo "L4T_VERSION: $L4T_VERSION"
|
33
|
+
if [[ ! " ${ALLOWED_L4T_VERSIONS[@]} " =~ " ${L4T_VERSION} " ]]; then
|
34
|
+
echo "L4T_VERSION is not in the allowed versions list. Exiting."
|
35
|
+
exit 1
|
36
|
+
fi
|
37
|
+
|
38
|
+
|
39
|
+
BASE_PATH=/home/$USER/reComputer
|
40
|
+
mkdir -p $BASE_PATH/
|
41
|
+
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
42
|
+
BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
|
43
|
+
if [ -d $JETSON_REPO_PATH ]; then
|
44
|
+
echo "jetson-ai-lab existed."
|
45
|
+
else
|
46
|
+
echo "jetson-ai-lab does not installed. start init..."
|
47
|
+
cd $BASE_PATH/
|
48
|
+
git clone --depth=1 $BASE_JETSON_LAB_GIT
|
49
|
+
cd $JETSON_REPO_PATH
|
50
|
+
bash install.sh
|
51
|
+
fi
|
52
|
+
|
@@ -1,10 +1,10 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
|
4
|
-
DATA_PATH="/home/$USER/reComputer/jetson-containers/data"
|
5
|
-
|
6
|
-
sudo docker run -it --rm --network host --runtime nvidia \
|
7
|
-
--volume $DATA_PATH:/data \
|
8
|
-
--name llama-factory \
|
9
|
-
youjiang9977/llama-factory:r35.4.1
|
10
|
-
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
|
4
|
+
DATA_PATH="/home/$USER/reComputer/jetson-containers/data"
|
5
|
+
|
6
|
+
sudo docker run -it --rm --network host --runtime nvidia \
|
7
|
+
--volume $DATA_PATH:/data \
|
8
|
+
--name llama-factory \
|
9
|
+
youjiang9977/llama-factory:r35.4.1
|
10
|
+
|
@@ -1,22 +1,22 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
BASE_PATH=/home/$USER/reComputer
|
3
|
-
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
4
|
-
# search local image
|
5
|
-
img_tag=$($JETSON_REPO_PATH/autotag -p local ollama)
|
6
|
-
# 检查返回值
|
7
|
-
if [ $? -eq 0 ]; then
|
8
|
-
echo "Found Image successfully."
|
9
|
-
sudo docker rmi $img_tag
|
10
|
-
else
|
11
|
-
echo "[warn] Found Image failed with error code $?. skip delete Image."
|
12
|
-
fi
|
13
|
-
#
|
14
|
-
# 4 build whl
|
15
|
-
read -p "Delete all data for ollama? (y/n): " choice
|
16
|
-
if [[ $choice == "y" || $choice == "Y" ]]; then
|
17
|
-
echo "Delete=> $JETSON_REPO_PATH/data/models/ollama/"
|
18
|
-
sudo rm -rf $JETSON_REPO_PATH/data/models/ollama/
|
19
|
-
echo "Clean Data Done."
|
20
|
-
else
|
21
|
-
echo "[warn] Skip Clean Data."
|
22
|
-
fi
|
1
|
+
#!/bin/bash
|
2
|
+
BASE_PATH=/home/$USER/reComputer
|
3
|
+
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
4
|
+
# search local image
|
5
|
+
img_tag=$($JETSON_REPO_PATH/autotag -p local ollama)
|
6
|
+
# 检查返回值
|
7
|
+
if [ $? -eq 0 ]; then
|
8
|
+
echo "Found Image successfully."
|
9
|
+
sudo docker rmi $img_tag
|
10
|
+
else
|
11
|
+
echo "[warn] Found Image failed with error code $?. skip delete Image."
|
12
|
+
fi
|
13
|
+
#
|
14
|
+
# 4 build whl
|
15
|
+
read -p "Delete all data for ollama? (y/n): " choice
|
16
|
+
if [[ $choice == "y" || $choice == "Y" ]]; then
|
17
|
+
echo "Delete=> $JETSON_REPO_PATH/data/models/ollama/"
|
18
|
+
sudo rm -rf $JETSON_REPO_PATH/data/models/ollama/
|
19
|
+
echo "Clean Data Done."
|
20
|
+
else
|
21
|
+
echo "[warn] Skip Clean Data."
|
22
|
+
fi
|
@@ -0,0 +1,31 @@
|
|
1
|
+
# The tested JetPack versions.
|
2
|
+
ALLOWED_L4T_VERSIONS:
|
3
|
+
- 35.3.1
|
4
|
+
- 35.4.1
|
5
|
+
- 35.5.0
|
6
|
+
- 36.3.0
|
7
|
+
REQUIRED_DISK_SPACE: 15 # in GB
|
8
|
+
REQUIRED_MEM_SPACE: 7
|
9
|
+
PACKAGES:
|
10
|
+
- nvidia-jetpack
|
11
|
+
DOCKER:
|
12
|
+
ENABLE: true
|
13
|
+
DAEMON: |
|
14
|
+
{
|
15
|
+
"default-runtime": "nvidia",
|
16
|
+
"runtimes": {
|
17
|
+
"nvidia": {
|
18
|
+
"path": "nvidia-container-runtime",
|
19
|
+
"runtimeArgs": []
|
20
|
+
}
|
21
|
+
},
|
22
|
+
"storage-driver": "overlay2",
|
23
|
+
"data-root": "/var/lib/docker",
|
24
|
+
"log-driver": "json-file",
|
25
|
+
"log-opts": {
|
26
|
+
"max-size": "100m",
|
27
|
+
"max-file": "3"
|
28
|
+
},
|
29
|
+
"no-new-privileges": true,
|
30
|
+
"experimental": false
|
31
|
+
}
|
@@ -1,16 +1,19 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
|
4
|
-
|
5
|
-
|
6
|
-
|
7
|
-
|
8
|
-
|
9
|
-
|
10
|
-
|
11
|
-
|
12
|
-
|
13
|
-
|
14
|
-
|
15
|
-
|
16
|
-
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
# check the runtime environment.
|
4
|
+
source $(dirname "$(realpath "$0")")/../utils.sh
|
5
|
+
check_base_env "$(dirname "$(realpath "$0")")/config.yaml"
|
6
|
+
|
7
|
+
BASE_PATH=/home/$USER/reComputer
|
8
|
+
mkdir -p $BASE_PATH/
|
9
|
+
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
10
|
+
BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
|
11
|
+
if [ -d $JETSON_REPO_PATH ]; then
|
12
|
+
echo "jetson-ai-lab existed."
|
13
|
+
else
|
14
|
+
echo "jetson-ai-lab does not installed. start init..."
|
15
|
+
cd $BASE_PATH/
|
16
|
+
git clone --depth=1 $BASE_JETSON_LAB_GIT
|
17
|
+
cd $JETSON_REPO_PATH
|
18
|
+
bash install.sh
|
19
|
+
fi
|
reComputer/scripts/llama3/run.sh
CHANGED
@@ -1,13 +1,13 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
BASE_PATH=/home/$USER/reComputer
|
4
|
-
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
5
|
-
cd $JETSON_REPO_PATH
|
6
|
-
# try stop old server
|
7
|
-
docker rm -f ollama
|
8
|
-
# start new server
|
9
|
-
./run.sh -d --name ollama $(./autotag ollama)
|
10
|
-
# run a client
|
11
|
-
./run.sh $(./autotag ollama) /bin/ollama run llama3
|
12
|
-
# clean new server
|
13
|
-
docker rm -f ollama
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
BASE_PATH=/home/$USER/reComputer
|
4
|
+
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
5
|
+
cd $JETSON_REPO_PATH
|
6
|
+
# try stop old server
|
7
|
+
docker rm -f ollama
|
8
|
+
# start new server
|
9
|
+
./run.sh -d --name ollama $(./autotag ollama)
|
10
|
+
# run a client
|
11
|
+
./run.sh $(./autotag ollama) /bin/ollama run llama3
|
12
|
+
# clean new server
|
13
|
+
docker rm -f ollama
|
@@ -1,4 +1,4 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
# TODO: clean old container
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
# TODO: clean old container
|
4
4
|
docker rmi $(./autotag llava)
|
reComputer/scripts/llava/init.sh
CHANGED
@@ -1,16 +1,16 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
|
4
|
-
BASE_PATH=/home/$USER/reComputer
|
5
|
-
mkdir -p $BASE_PATH/
|
6
|
-
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
7
|
-
BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
|
8
|
-
if [ -d $JETSON_REPO_PATH ]; then
|
9
|
-
echo "jetson-ai-lab existed."
|
10
|
-
else
|
11
|
-
echo "jetson-ai-lab does not installed. start init..."
|
12
|
-
cd $BASE_PATH/
|
13
|
-
git clone --depth=1 $BASE_JETSON_LAB_GIT
|
14
|
-
cd $JETSON_REPO_PATH
|
15
|
-
bash install.sh
|
16
|
-
fi
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
|
4
|
+
BASE_PATH=/home/$USER/reComputer
|
5
|
+
mkdir -p $BASE_PATH/
|
6
|
+
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
7
|
+
BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
|
8
|
+
if [ -d $JETSON_REPO_PATH ]; then
|
9
|
+
echo "jetson-ai-lab existed."
|
10
|
+
else
|
11
|
+
echo "jetson-ai-lab does not installed. start init..."
|
12
|
+
cd $BASE_PATH/
|
13
|
+
git clone --depth=1 $BASE_JETSON_LAB_GIT
|
14
|
+
cd $JETSON_REPO_PATH
|
15
|
+
bash install.sh
|
16
|
+
fi
|
reComputer/scripts/llava/run.sh
CHANGED
@@ -1,10 +1,10 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
BASE_PATH=/home/$USER/reComputer
|
4
|
-
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
5
|
-
cd $JETSON_REPO_PATH
|
6
|
-
|
7
|
-
./run.sh $(./autotag llava) \
|
8
|
-
python3 -m llava.serve.cli \
|
9
|
-
--model-path liuhaotian/llava-v1.5-7b \
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
BASE_PATH=/home/$USER/reComputer
|
4
|
+
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
5
|
+
cd $JETSON_REPO_PATH
|
6
|
+
|
7
|
+
./run.sh $(./autotag llava) \
|
8
|
+
python3 -m llava.serve.cli \
|
9
|
+
--model-path liuhaotian/llava-v1.5-7b \
|
10
10
|
--image-file /data/images/hoover.jpg
|
@@ -1,16 +1,16 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
|
4
|
-
BASE_PATH=/home/$USER/reComputer
|
5
|
-
mkdir -p $BASE_PATH/
|
6
|
-
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
7
|
-
BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
|
8
|
-
if [ -d $JETSON_REPO_PATH ]; then
|
9
|
-
echo "jetson-ai-lab existed."
|
10
|
-
else
|
11
|
-
echo "jetson-ai-lab does not installed. start init..."
|
12
|
-
cd $BASE_PATH/
|
13
|
-
git clone --depth=1 $BASE_JETSON_LAB_GIT
|
14
|
-
cd $JETSON_REPO_PATH
|
15
|
-
bash install.sh
|
16
|
-
fi
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
|
4
|
+
BASE_PATH=/home/$USER/reComputer
|
5
|
+
mkdir -p $BASE_PATH/
|
6
|
+
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
7
|
+
BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
|
8
|
+
if [ -d $JETSON_REPO_PATH ]; then
|
9
|
+
echo "jetson-ai-lab existed."
|
10
|
+
else
|
11
|
+
echo "jetson-ai-lab does not installed. start init..."
|
12
|
+
cd $BASE_PATH/
|
13
|
+
git clone --depth=1 $BASE_JETSON_LAB_GIT
|
14
|
+
cd $JETSON_REPO_PATH
|
15
|
+
bash install.sh
|
16
|
+
fi
|
@@ -1,10 +1,10 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
BASE_PATH=/home/$USER/reComputer
|
4
|
-
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
5
|
-
cd $JETSON_REPO_PATH
|
6
|
-
|
7
|
-
./run.sh $(./autotag llava) \
|
8
|
-
python3 -m llava.serve.cli \
|
9
|
-
--model-path liuhaotian/llava-v1.5-7b \
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
BASE_PATH=/home/$USER/reComputer
|
4
|
+
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
5
|
+
cd $JETSON_REPO_PATH
|
6
|
+
|
7
|
+
./run.sh $(./autotag llava) \
|
8
|
+
python3 -m llava.serve.cli \
|
9
|
+
--model-path liuhaotian/llava-v1.5-7b \
|
10
10
|
--image-file /data/images/hoover.jpg
|
@@ -1,16 +1,16 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
|
4
|
-
BASE_PATH=/home/$USER/reComputer
|
5
|
-
mkdir -p $BASE_PATH/
|
6
|
-
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
7
|
-
BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
|
8
|
-
if [ -d $JETSON_REPO_PATH ]; then
|
9
|
-
echo "jetson-ai-lab existed."
|
10
|
-
else
|
11
|
-
echo "jetson-ai-lab does not installed. start init..."
|
12
|
-
cd $BASE_PATH/
|
13
|
-
git clone --depth=1 $BASE_JETSON_LAB_GIT
|
14
|
-
cd $JETSON_REPO_PATH
|
15
|
-
bash install.sh
|
16
|
-
fi
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
|
4
|
+
BASE_PATH=/home/$USER/reComputer
|
5
|
+
mkdir -p $BASE_PATH/
|
6
|
+
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
7
|
+
BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
|
8
|
+
if [ -d $JETSON_REPO_PATH ]; then
|
9
|
+
echo "jetson-ai-lab existed."
|
10
|
+
else
|
11
|
+
echo "jetson-ai-lab does not installed. start init..."
|
12
|
+
cd $BASE_PATH/
|
13
|
+
git clone --depth=1 $BASE_JETSON_LAB_GIT
|
14
|
+
cd $JETSON_REPO_PATH
|
15
|
+
bash install.sh
|
16
|
+
fi
|
@@ -1,11 +1,11 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
BASE_PATH=/home/$USER/reComputer
|
4
|
-
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
5
|
-
cd $JETSON_REPO_PATH
|
6
|
-
|
7
|
-
./run.sh $(./autotag local_llm) \
|
8
|
-
python3 -m local_llm --api=mlc \
|
9
|
-
--model liuhaotian/llava-v1.6-vicuna-7b \
|
10
|
-
--max-context-len 768 \
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
BASE_PATH=/home/$USER/reComputer
|
4
|
+
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
5
|
+
cd $JETSON_REPO_PATH
|
6
|
+
|
7
|
+
./run.sh $(./autotag local_llm) \
|
8
|
+
python3 -m local_llm --api=mlc \
|
9
|
+
--model liuhaotian/llava-v1.6-vicuna-7b \
|
10
|
+
--max-context-len 768 \
|
11
11
|
--max-new-tokens 128
|
@@ -1,16 +1,16 @@
|
|
1
|
-
#!/bin/bash
|
2
|
-
|
3
|
-
|
4
|
-
BASE_PATH=/home/$USER/reComputer
|
5
|
-
mkdir -p $BASE_PATH/
|
6
|
-
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
7
|
-
BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
|
8
|
-
if [ -d $JETSON_REPO_PATH ]; then
|
9
|
-
echo "jetson-ai-lab existed."
|
10
|
-
else
|
11
|
-
echo "jetson-ai-lab does not installed. start init..."
|
12
|
-
cd $BASE_PATH/
|
13
|
-
git clone --depth=1 $BASE_JETSON_LAB_GIT
|
14
|
-
cd $JETSON_REPO_PATH
|
15
|
-
bash install.sh
|
16
|
-
fi
|
1
|
+
#!/bin/bash
|
2
|
+
|
3
|
+
|
4
|
+
BASE_PATH=/home/$USER/reComputer
|
5
|
+
mkdir -p $BASE_PATH/
|
6
|
+
JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
|
7
|
+
BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
|
8
|
+
if [ -d $JETSON_REPO_PATH ]; then
|
9
|
+
echo "jetson-ai-lab existed."
|
10
|
+
else
|
11
|
+
echo "jetson-ai-lab does not installed. start init..."
|
12
|
+
cd $BASE_PATH/
|
13
|
+
git clone --depth=1 $BASE_JETSON_LAB_GIT
|
14
|
+
cd $JETSON_REPO_PATH
|
15
|
+
bash install.sh
|
16
|
+
fi
|
@@ -1,10 +1,10 @@
|
|
1
|
-
# NanoDB
|
2
|
-
|
3
|
-
## ref
|
4
|
-
|
5
|
-
- <https://www.jetson-ai-lab.com/tutorial_nanodb.html>
|
6
|
-
|
7
|
-
## access
|
8
|
-
|
9
|
-
- using in machine, try `http://127.0.0.1:7860` in browser.
|
10
|
-
- using in other pc, make sure you know jetson's IP and try `http://<<Jetson's IP>>:7860` in browser.
|
1
|
+
# NanoDB
|
2
|
+
|
3
|
+
## ref
|
4
|
+
|
5
|
+
- <https://www.jetson-ai-lab.com/tutorial_nanodb.html>
|
6
|
+
|
7
|
+
## access
|
8
|
+
|
9
|
+
- using in machine, try `http://127.0.0.1:7860` in browser.
|
10
|
+
- using in other pc, make sure you know jetson's IP and try `http://<<Jetson's IP>>:7860` in browser.
|