jetson-examples 0.1.8__py3-none-any.whl → 0.1.9__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (106) hide show
  1. {jetson_examples-0.1.8.dist-info → jetson_examples-0.1.9.dist-info}/LICENSE +21 -21
  2. {jetson_examples-0.1.8.dist-info → jetson_examples-0.1.9.dist-info}/METADATA +1 -1
  3. jetson_examples-0.1.9.dist-info/RECORD +109 -0
  4. reComputer/__init__.py +1 -1
  5. reComputer/main.py +60 -60
  6. reComputer/scripts/MoveNet-Lightning/clean.sh +8 -8
  7. reComputer/scripts/MoveNet-Lightning/getVersion.sh +59 -59
  8. reComputer/scripts/MoveNet-Lightning/init.sh +6 -6
  9. reComputer/scripts/MoveNet-Lightning/readme.md +30 -30
  10. reComputer/scripts/MoveNet-Lightning/run.sh +19 -19
  11. reComputer/scripts/MoveNet-Thunder/clean.sh +7 -7
  12. reComputer/scripts/MoveNet-Thunder/getVersion.sh +59 -59
  13. reComputer/scripts/MoveNet-Thunder/init.sh +6 -6
  14. reComputer/scripts/MoveNet-Thunder/readme.md +31 -31
  15. reComputer/scripts/MoveNet-Thunder/run.sh +18 -18
  16. reComputer/scripts/MoveNetJS/clean.sh +4 -4
  17. reComputer/scripts/MoveNetJS/readme.md +56 -56
  18. reComputer/scripts/MoveNetJS/run.sh +13 -13
  19. reComputer/scripts/Sheared-LLaMA-2.7B-ShareGPT/init.sh +16 -16
  20. reComputer/scripts/Sheared-LLaMA-2.7B-ShareGPT/run.sh +8 -8
  21. reComputer/scripts/audiocraft/README.md +35 -35
  22. reComputer/scripts/audiocraft/clean.sh +5 -5
  23. reComputer/scripts/audiocraft/init.sh +16 -16
  24. reComputer/scripts/audiocraft/run.sh +7 -7
  25. reComputer/scripts/check.sh +4 -4
  26. reComputer/scripts/clean.sh +33 -33
  27. reComputer/scripts/comfyui/LICENSE +21 -21
  28. reComputer/scripts/comfyui/README.md +127 -127
  29. reComputer/scripts/comfyui/clean.sh +9 -7
  30. reComputer/scripts/comfyui/config.yaml +30 -29
  31. reComputer/scripts/comfyui/init.sh +9 -163
  32. reComputer/scripts/comfyui/run.sh +30 -30
  33. reComputer/scripts/depth-anything/Dockerfile +5 -5
  34. reComputer/scripts/depth-anything/LICENSE +21 -21
  35. reComputer/scripts/depth-anything/README.md +135 -135
  36. reComputer/scripts/depth-anything/clean.sh +7 -7
  37. reComputer/scripts/depth-anything/config.yaml +31 -31
  38. reComputer/scripts/depth-anything/init.sh +164 -164
  39. reComputer/scripts/depth-anything/run.sh +22 -22
  40. reComputer/scripts/depth-anything-v2/Dockerfile +5 -5
  41. reComputer/scripts/depth-anything-v2/LICENSE +21 -21
  42. reComputer/scripts/depth-anything-v2/README.md +135 -135
  43. reComputer/scripts/depth-anything-v2/clean.sh +7 -7
  44. reComputer/scripts/depth-anything-v2/config.yaml +31 -31
  45. reComputer/scripts/depth-anything-v2/init.sh +164 -164
  46. reComputer/scripts/depth-anything-v2/run.sh +22 -22
  47. reComputer/scripts/live-llava/init.sh +16 -16
  48. reComputer/scripts/live-llava/run.sh +278 -278
  49. reComputer/scripts/llama-factory/README.md +68 -68
  50. reComputer/scripts/llama-factory/clean.sh +4 -4
  51. reComputer/scripts/llama-factory/init.sh +52 -52
  52. reComputer/scripts/llama-factory/run.sh +10 -10
  53. reComputer/scripts/llama3/clean.sh +22 -22
  54. reComputer/scripts/llama3/config.yaml +31 -0
  55. reComputer/scripts/llama3/init.sh +19 -16
  56. reComputer/scripts/llama3/run.sh +13 -13
  57. reComputer/scripts/llava/clean.sh +3 -3
  58. reComputer/scripts/llava/init.sh +16 -16
  59. reComputer/scripts/llava/run.sh +9 -9
  60. reComputer/scripts/llava-v1.5-7b/init.sh +16 -16
  61. reComputer/scripts/llava-v1.5-7b/run.sh +9 -9
  62. reComputer/scripts/llava-v1.6-vicuna-7b/init.sh +16 -16
  63. reComputer/scripts/llava-v1.6-vicuna-7b/run.sh +10 -10
  64. reComputer/scripts/nanodb/init.sh +16 -16
  65. reComputer/scripts/nanodb/readme.md +10 -10
  66. reComputer/scripts/nanodb/run.sh +90 -90
  67. reComputer/scripts/nanoowl/init.sh +16 -16
  68. reComputer/scripts/nanoowl/run.sh +7 -7
  69. reComputer/scripts/ollama/clean.sh +22 -22
  70. reComputer/scripts/ollama/config.yaml +31 -0
  71. reComputer/scripts/ollama/init.sh +19 -16
  72. reComputer/scripts/ollama/run.sh +10 -10
  73. reComputer/scripts/parler-tts/clean.sh +7 -7
  74. reComputer/scripts/parler-tts/getVersion.sh +59 -59
  75. reComputer/scripts/parler-tts/init.sh +8 -8
  76. reComputer/scripts/parler-tts/readme.md +63 -63
  77. reComputer/scripts/parler-tts/run.sh +17 -17
  78. reComputer/scripts/run.sh +48 -48
  79. reComputer/scripts/stable-diffusion-webui/init.sh +16 -16
  80. reComputer/scripts/stable-diffusion-webui/run.sh +6 -6
  81. reComputer/scripts/text-generation-webui/init.sh +16 -16
  82. reComputer/scripts/text-generation-webui/run.sh +11 -11
  83. reComputer/scripts/ultralytics-yolo/LICENSE +21 -21
  84. reComputer/scripts/ultralytics-yolo/README.md +124 -124
  85. reComputer/scripts/ultralytics-yolo/clean.sh +6 -6
  86. reComputer/scripts/ultralytics-yolo/config.yaml +31 -31
  87. reComputer/scripts/ultralytics-yolo/init.sh +4 -4
  88. reComputer/scripts/ultralytics-yolo/run.sh +26 -26
  89. reComputer/scripts/update.sh +26 -26
  90. reComputer/scripts/utils.sh +168 -166
  91. reComputer/scripts/whisper/init.sh +16 -16
  92. reComputer/scripts/whisper/run.sh +7 -7
  93. reComputer/scripts/yolov10/Dockerfile +13 -13
  94. reComputer/scripts/yolov10/README.md +71 -71
  95. reComputer/scripts/yolov10/clean.sh +4 -4
  96. reComputer/scripts/yolov10/config.yaml +31 -31
  97. reComputer/scripts/yolov10/init.sh +20 -20
  98. reComputer/scripts/yolov10/run.sh +7 -7
  99. reComputer/scripts/yolov8-rail-inspection/config.yaml +31 -31
  100. reComputer/scripts/yolov8-rail-inspection/init.sh +5 -5
  101. reComputer/scripts/yolov8-rail-inspection/readme.md +35 -35
  102. reComputer/scripts/yolov8-rail-inspection/run.sh +21 -21
  103. jetson_examples-0.1.8.dist-info/RECORD +0 -107
  104. {jetson_examples-0.1.8.dist-info → jetson_examples-0.1.9.dist-info}/WHEEL +0 -0
  105. {jetson_examples-0.1.8.dist-info → jetson_examples-0.1.9.dist-info}/entry_points.txt +0 -0
  106. {jetson_examples-0.1.8.dist-info → jetson_examples-0.1.9.dist-info}/top_level.txt +0 -0
@@ -1,68 +1,68 @@
1
- # Finetune LLM by Llama-Factory on Jetson
2
-
3
-
4
- ## Hello
5
- Now you can tailor a custom private local LLM to meet your requirements.
6
-
7
- 💡 Here's an example of quickly deploying [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) on Jetson device.
8
-
9
- 🔥 Hightlights:
10
- - **Llama-Factory** is an efficient tool to unify efficient Fine-Tuning of 100+ LLMs. 🚀🔍
11
- - **Jetson-examples** is a toolkit designed to deploy containerized applications on NVIDIA Jetson devices. ✨
12
- - **Jetson** is powerful AI hardware platform for edge computing.💻
13
-
14
- 🛠️ Follow the tutorial below to quickly experience the performance of Llama-Factory on edge computing devices.
15
-
16
- <div align="center">
17
- <img alt="training" width="1200px" src="./assets/training.gif">
18
- </div>
19
-
20
- ## Get a Jetson Orin Device 🛒
21
- | Device Model | Description | Link |
22
- |--------------|-------------|------|
23
- | reComputer J4012, powered by Orin NX 16GB, 100 TOPS | Embedded computer powered by Orin NX | [Buy Here](https://www.seeedstudio.com/reComputer-J4012-p-5586.html) |
24
- | NVIDIA® Jetson AGX Orin™ 64GB Developer Kit | Smallest and most powerful AI edge computer | [Buy Here](https://www.seeedstudio.com/NVIDIArJetson-AGX-Orintm-64GB-Developer-Kit-p-5641.html) |
25
-
26
- ## Getting Started
27
-
28
- - install **jetson-examples** by pip:
29
- ```sh
30
- pip3 install jetson-examples
31
- ```
32
- - restart reComputer
33
- ```sh
34
- sudo restart
35
- ```
36
- - run Llama-Factory webui on jetson in one line:
37
- ```sh
38
- reComputer run llama-factory
39
- ```
40
- - Please visit http://127.0.0.1:7860
41
-
42
- <div align="center">
43
- <img alt="yolov10" width="1200px" src="./assets/webui.png">
44
- </div>
45
-
46
-
47
-
48
- ## Run Training Script
49
-
50
- > **Note:** Some models and datasets require confirmation before using them, so we recommend logging in with your Hugging Face account by:
51
- > `sudo docker exec -it llama-factory huggingface-cli login`
52
-
53
- There are a lot of parameters to choose from webui, you can refer to here for more information.
54
-
55
- For demonstration purposes, set `Model name: Phi-1.5-1.3B`, `Dataset: alpaca_zh`, leave the other parameters unchanged, and then click the `Start` button
56
-
57
- <div align="center">
58
- <img alt="yolov10" width="1200px" src="./assets/llama-factory-Jetson.png">
59
- </div>
60
-
61
-
62
- ## Build Docker Image
63
- We highly recommend that you use `jetson-containers` to compile the docker container, as you can see [here](https://github.com/dusty-nv/jetson-containers/pull/566).
64
-
65
- ## Reference
66
- - https://github.com/hiyouga/LLaMA-Factory
67
- - https://github.com/dusty-nv/jetson-containers
68
-
1
+ # Finetune LLM by Llama-Factory on Jetson
2
+
3
+
4
+ ## Hello
5
+ Now you can tailor a custom private local LLM to meet your requirements.
6
+
7
+ 💡 Here's an example of quickly deploying [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) on Jetson device.
8
+
9
+ 🔥 Hightlights:
10
+ - **Llama-Factory** is an efficient tool to unify efficient Fine-Tuning of 100+ LLMs. 🚀🔍
11
+ - **Jetson-examples** is a toolkit designed to deploy containerized applications on NVIDIA Jetson devices. ✨
12
+ - **Jetson** is powerful AI hardware platform for edge computing.💻
13
+
14
+ 🛠️ Follow the tutorial below to quickly experience the performance of Llama-Factory on edge computing devices.
15
+
16
+ <div align="center">
17
+ <img alt="training" width="1200px" src="./assets/training.gif">
18
+ </div>
19
+
20
+ ## Get a Jetson Orin Device 🛒
21
+ | Device Model | Description | Link |
22
+ |--------------|-------------|------|
23
+ | reComputer J4012, powered by Orin NX 16GB, 100 TOPS | Embedded computer powered by Orin NX | [Buy Here](https://www.seeedstudio.com/reComputer-J4012-p-5586.html) |
24
+ | NVIDIA® Jetson AGX Orin™ 64GB Developer Kit | Smallest and most powerful AI edge computer | [Buy Here](https://www.seeedstudio.com/NVIDIArJetson-AGX-Orintm-64GB-Developer-Kit-p-5641.html) |
25
+
26
+ ## Getting Started
27
+
28
+ - install **jetson-examples** by pip:
29
+ ```sh
30
+ pip3 install jetson-examples
31
+ ```
32
+ - restart reComputer
33
+ ```sh
34
+ sudo restart
35
+ ```
36
+ - run Llama-Factory webui on jetson in one line:
37
+ ```sh
38
+ reComputer run llama-factory
39
+ ```
40
+ - Please visit http://127.0.0.1:7860
41
+
42
+ <div align="center">
43
+ <img alt="yolov10" width="1200px" src="./assets/webui.png">
44
+ </div>
45
+
46
+
47
+
48
+ ## Run Training Script
49
+
50
+ > **Note:** Some models and datasets require confirmation before using them, so we recommend logging in with your Hugging Face account by:
51
+ > `sudo docker exec -it llama-factory huggingface-cli login`
52
+
53
+ There are a lot of parameters to choose from webui, you can refer to here for more information.
54
+
55
+ For demonstration purposes, set `Model name: Phi-1.5-1.3B`, `Dataset: alpaca_zh`, leave the other parameters unchanged, and then click the `Start` button
56
+
57
+ <div align="center">
58
+ <img alt="yolov10" width="1200px" src="./assets/llama-factory-Jetson.png">
59
+ </div>
60
+
61
+
62
+ ## Build Docker Image
63
+ We highly recommend that you use `jetson-containers` to compile the docker container, as you can see [here](https://github.com/dusty-nv/jetson-containers/pull/566).
64
+
65
+ ## Reference
66
+ - https://github.com/hiyouga/LLaMA-Factory
67
+ - https://github.com/dusty-nv/jetson-containers
68
+
@@ -1,4 +1,4 @@
1
- #!/bin/bash
2
-
3
- sudo docker rmi youjiang9977/llama-factory:r35.4.1
4
- sudo rm -rf /home/$USER/reComputer/jetson-containers/LLaMA-Factory/*
1
+ #!/bin/bash
2
+
3
+ sudo docker rmi youjiang9977/llama-factory:r35.4.1
4
+ sudo rm -rf /home/$USER/reComputer/jetson-containers/LLaMA-Factory/*
@@ -1,52 +1,52 @@
1
- #!/bin/bash
2
-
3
-
4
- # check L4T_VERSION
5
- # reference https://github.com/dusty-nv/jetson-containers/blob/master/jetson_containers/l4t_version.sh
6
- ALLOWED_L4T_VERSIONS=("35.3.1" "35.4.1" "35.5.0")
7
-
8
- ARCH=$(uname -i)
9
- if [ $ARCH = "aarch64" ]; then
10
- L4T_VERSION_STRING=$(head -n 1 /etc/nv_tegra_release)
11
-
12
- if [ -z "$L4T_VERSION_STRING" ]; then
13
- L4T_VERSION_STRING=$(dpkg-query --showformat='${Version}' --show nvidia-l4t-core)
14
- L4T_VERSION_ARRAY=(${L4T_VERSION_STRING//./ })
15
- L4T_RELEASE=${L4T_VERSION_ARRAY[0]}
16
- L4T_REVISION=${L4T_VERSION_ARRAY[1]}
17
- else
18
- L4T_RELEASE=$(echo $L4T_VERSION_STRING | cut -f 2 -d ' ' | grep -Po '(?<=R)[^;]+')
19
- L4T_REVISION=$(echo $L4T_VERSION_STRING | cut -f 2 -d ',' | grep -Po '(?<=REVISION: )[^;]+')
20
- fi
21
-
22
- L4T_REVISION_MAJOR=${L4T_REVISION:0:1}
23
- L4T_REVISION_MINOR=${L4T_REVISION:2:1}
24
-
25
- L4T_VERSION="$L4T_RELEASE.$L4T_REVISION"
26
- echo $L4T_VERSION
27
- elif [ $ARCH != "x86_64" ]; then
28
- echo "unsupported architecture: $ARCH"
29
- exit 1
30
- fi
31
-
32
- echo "L4T_VERSION: $L4T_VERSION"
33
- if [[ ! " ${ALLOWED_L4T_VERSIONS[@]} " =~ " ${L4T_VERSION} " ]]; then
34
- echo "L4T_VERSION is not in the allowed versions list. Exiting."
35
- exit 1
36
- fi
37
-
38
-
39
- BASE_PATH=/home/$USER/reComputer
40
- mkdir -p $BASE_PATH/
41
- JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
42
- BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
43
- if [ -d $JETSON_REPO_PATH ]; then
44
- echo "jetson-ai-lab existed."
45
- else
46
- echo "jetson-ai-lab does not installed. start init..."
47
- cd $BASE_PATH/
48
- git clone --depth=1 $BASE_JETSON_LAB_GIT
49
- cd $JETSON_REPO_PATH
50
- bash install.sh
51
- fi
52
-
1
+ #!/bin/bash
2
+
3
+
4
+ # check L4T_VERSION
5
+ # reference https://github.com/dusty-nv/jetson-containers/blob/master/jetson_containers/l4t_version.sh
6
+ ALLOWED_L4T_VERSIONS=("35.3.1" "35.4.1" "35.5.0")
7
+
8
+ ARCH=$(uname -i)
9
+ if [ $ARCH = "aarch64" ]; then
10
+ L4T_VERSION_STRING=$(head -n 1 /etc/nv_tegra_release)
11
+
12
+ if [ -z "$L4T_VERSION_STRING" ]; then
13
+ L4T_VERSION_STRING=$(dpkg-query --showformat='${Version}' --show nvidia-l4t-core)
14
+ L4T_VERSION_ARRAY=(${L4T_VERSION_STRING//./ })
15
+ L4T_RELEASE=${L4T_VERSION_ARRAY[0]}
16
+ L4T_REVISION=${L4T_VERSION_ARRAY[1]}
17
+ else
18
+ L4T_RELEASE=$(echo $L4T_VERSION_STRING | cut -f 2 -d ' ' | grep -Po '(?<=R)[^;]+')
19
+ L4T_REVISION=$(echo $L4T_VERSION_STRING | cut -f 2 -d ',' | grep -Po '(?<=REVISION: )[^;]+')
20
+ fi
21
+
22
+ L4T_REVISION_MAJOR=${L4T_REVISION:0:1}
23
+ L4T_REVISION_MINOR=${L4T_REVISION:2:1}
24
+
25
+ L4T_VERSION="$L4T_RELEASE.$L4T_REVISION"
26
+ echo $L4T_VERSION
27
+ elif [ $ARCH != "x86_64" ]; then
28
+ echo "unsupported architecture: $ARCH"
29
+ exit 1
30
+ fi
31
+
32
+ echo "L4T_VERSION: $L4T_VERSION"
33
+ if [[ ! " ${ALLOWED_L4T_VERSIONS[@]} " =~ " ${L4T_VERSION} " ]]; then
34
+ echo "L4T_VERSION is not in the allowed versions list. Exiting."
35
+ exit 1
36
+ fi
37
+
38
+
39
+ BASE_PATH=/home/$USER/reComputer
40
+ mkdir -p $BASE_PATH/
41
+ JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
42
+ BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
43
+ if [ -d $JETSON_REPO_PATH ]; then
44
+ echo "jetson-ai-lab existed."
45
+ else
46
+ echo "jetson-ai-lab does not installed. start init..."
47
+ cd $BASE_PATH/
48
+ git clone --depth=1 $BASE_JETSON_LAB_GIT
49
+ cd $JETSON_REPO_PATH
50
+ bash install.sh
51
+ fi
52
+
@@ -1,10 +1,10 @@
1
- #!/bin/bash
2
-
3
-
4
- DATA_PATH="/home/$USER/reComputer/jetson-containers/data"
5
-
6
- sudo docker run -it --rm --network host --runtime nvidia \
7
- --volume $DATA_PATH:/data \
8
- --name llama-factory \
9
- youjiang9977/llama-factory:r35.4.1
10
-
1
+ #!/bin/bash
2
+
3
+
4
+ DATA_PATH="/home/$USER/reComputer/jetson-containers/data"
5
+
6
+ sudo docker run -it --rm --network host --runtime nvidia \
7
+ --volume $DATA_PATH:/data \
8
+ --name llama-factory \
9
+ youjiang9977/llama-factory:r35.4.1
10
+
@@ -1,22 +1,22 @@
1
- #!/bin/bash
2
- BASE_PATH=/home/$USER/reComputer
3
- JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
4
- # search local image
5
- img_tag=$($JETSON_REPO_PATH/autotag -p local ollama)
6
- # 检查返回值
7
- if [ $? -eq 0 ]; then
8
- echo "Found Image successfully."
9
- sudo docker rmi $img_tag
10
- else
11
- echo "[warn] Found Image failed with error code $?. skip delete Image."
12
- fi
13
- #
14
- # 4 build whl
15
- read -p "Delete all data for ollama? (y/n): " choice
16
- if [[ $choice == "y" || $choice == "Y" ]]; then
17
- echo "Delete=> $JETSON_REPO_PATH/data/models/ollama/"
18
- sudo rm -rf $JETSON_REPO_PATH/data/models/ollama/
19
- echo "Clean Data Done."
20
- else
21
- echo "[warn] Skip Clean Data."
22
- fi
1
+ #!/bin/bash
2
+ BASE_PATH=/home/$USER/reComputer
3
+ JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
4
+ # search local image
5
+ img_tag=$($JETSON_REPO_PATH/autotag -p local ollama)
6
+ # 检查返回值
7
+ if [ $? -eq 0 ]; then
8
+ echo "Found Image successfully."
9
+ sudo docker rmi $img_tag
10
+ else
11
+ echo "[warn] Found Image failed with error code $?. skip delete Image."
12
+ fi
13
+ #
14
+ # 4 build whl
15
+ read -p "Delete all data for ollama? (y/n): " choice
16
+ if [[ $choice == "y" || $choice == "Y" ]]; then
17
+ echo "Delete=> $JETSON_REPO_PATH/data/models/ollama/"
18
+ sudo rm -rf $JETSON_REPO_PATH/data/models/ollama/
19
+ echo "Clean Data Done."
20
+ else
21
+ echo "[warn] Skip Clean Data."
22
+ fi
@@ -0,0 +1,31 @@
1
+ # The tested JetPack versions.
2
+ ALLOWED_L4T_VERSIONS:
3
+ - 35.3.1
4
+ - 35.4.1
5
+ - 35.5.0
6
+ - 36.3.0
7
+ REQUIRED_DISK_SPACE: 15 # in GB
8
+ REQUIRED_MEM_SPACE: 7
9
+ PACKAGES:
10
+ - nvidia-jetpack
11
+ DOCKER:
12
+ ENABLE: true
13
+ DAEMON: |
14
+ {
15
+ "default-runtime": "nvidia",
16
+ "runtimes": {
17
+ "nvidia": {
18
+ "path": "nvidia-container-runtime",
19
+ "runtimeArgs": []
20
+ }
21
+ },
22
+ "storage-driver": "overlay2",
23
+ "data-root": "/var/lib/docker",
24
+ "log-driver": "json-file",
25
+ "log-opts": {
26
+ "max-size": "100m",
27
+ "max-file": "3"
28
+ },
29
+ "no-new-privileges": true,
30
+ "experimental": false
31
+ }
@@ -1,16 +1,19 @@
1
- #!/bin/bash
2
-
3
-
4
- BASE_PATH=/home/$USER/reComputer
5
- mkdir -p $BASE_PATH/
6
- JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
7
- BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
8
- if [ -d $JETSON_REPO_PATH ]; then
9
- echo "jetson-ai-lab existed."
10
- else
11
- echo "jetson-ai-lab does not installed. start init..."
12
- cd $BASE_PATH/
13
- git clone --depth=1 $BASE_JETSON_LAB_GIT
14
- cd $JETSON_REPO_PATH
15
- bash install.sh
16
- fi
1
+ #!/bin/bash
2
+
3
+ # check the runtime environment.
4
+ source $(dirname "$(realpath "$0")")/../utils.sh
5
+ check_base_env "$(dirname "$(realpath "$0")")/config.yaml"
6
+
7
+ BASE_PATH=/home/$USER/reComputer
8
+ mkdir -p $BASE_PATH/
9
+ JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
10
+ BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
11
+ if [ -d $JETSON_REPO_PATH ]; then
12
+ echo "jetson-ai-lab existed."
13
+ else
14
+ echo "jetson-ai-lab does not installed. start init..."
15
+ cd $BASE_PATH/
16
+ git clone --depth=1 $BASE_JETSON_LAB_GIT
17
+ cd $JETSON_REPO_PATH
18
+ bash install.sh
19
+ fi
@@ -1,13 +1,13 @@
1
- #!/bin/bash
2
-
3
- BASE_PATH=/home/$USER/reComputer
4
- JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
5
- cd $JETSON_REPO_PATH
6
- # try stop old server
7
- docker rm -f ollama
8
- # start new server
9
- ./run.sh -d --name ollama $(./autotag ollama)
10
- # run a client
11
- ./run.sh $(./autotag ollama) /bin/ollama run llama3
12
- # clean new server
13
- docker rm -f ollama
1
+ #!/bin/bash
2
+
3
+ BASE_PATH=/home/$USER/reComputer
4
+ JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
5
+ cd $JETSON_REPO_PATH
6
+ # try stop old server
7
+ docker rm -f ollama
8
+ # start new server
9
+ ./run.sh -d --name ollama $(./autotag ollama)
10
+ # run a client
11
+ ./run.sh $(./autotag ollama) /bin/ollama run llama3
12
+ # clean new server
13
+ docker rm -f ollama
@@ -1,4 +1,4 @@
1
- #!/bin/bash
2
-
3
- # TODO: clean old container
1
+ #!/bin/bash
2
+
3
+ # TODO: clean old container
4
4
  docker rmi $(./autotag llava)
@@ -1,16 +1,16 @@
1
- #!/bin/bash
2
-
3
-
4
- BASE_PATH=/home/$USER/reComputer
5
- mkdir -p $BASE_PATH/
6
- JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
7
- BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
8
- if [ -d $JETSON_REPO_PATH ]; then
9
- echo "jetson-ai-lab existed."
10
- else
11
- echo "jetson-ai-lab does not installed. start init..."
12
- cd $BASE_PATH/
13
- git clone --depth=1 $BASE_JETSON_LAB_GIT
14
- cd $JETSON_REPO_PATH
15
- bash install.sh
16
- fi
1
+ #!/bin/bash
2
+
3
+
4
+ BASE_PATH=/home/$USER/reComputer
5
+ mkdir -p $BASE_PATH/
6
+ JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
7
+ BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
8
+ if [ -d $JETSON_REPO_PATH ]; then
9
+ echo "jetson-ai-lab existed."
10
+ else
11
+ echo "jetson-ai-lab does not installed. start init..."
12
+ cd $BASE_PATH/
13
+ git clone --depth=1 $BASE_JETSON_LAB_GIT
14
+ cd $JETSON_REPO_PATH
15
+ bash install.sh
16
+ fi
@@ -1,10 +1,10 @@
1
- #!/bin/bash
2
-
3
- BASE_PATH=/home/$USER/reComputer
4
- JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
5
- cd $JETSON_REPO_PATH
6
-
7
- ./run.sh $(./autotag llava) \
8
- python3 -m llava.serve.cli \
9
- --model-path liuhaotian/llava-v1.5-7b \
1
+ #!/bin/bash
2
+
3
+ BASE_PATH=/home/$USER/reComputer
4
+ JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
5
+ cd $JETSON_REPO_PATH
6
+
7
+ ./run.sh $(./autotag llava) \
8
+ python3 -m llava.serve.cli \
9
+ --model-path liuhaotian/llava-v1.5-7b \
10
10
  --image-file /data/images/hoover.jpg
@@ -1,16 +1,16 @@
1
- #!/bin/bash
2
-
3
-
4
- BASE_PATH=/home/$USER/reComputer
5
- mkdir -p $BASE_PATH/
6
- JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
7
- BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
8
- if [ -d $JETSON_REPO_PATH ]; then
9
- echo "jetson-ai-lab existed."
10
- else
11
- echo "jetson-ai-lab does not installed. start init..."
12
- cd $BASE_PATH/
13
- git clone --depth=1 $BASE_JETSON_LAB_GIT
14
- cd $JETSON_REPO_PATH
15
- bash install.sh
16
- fi
1
+ #!/bin/bash
2
+
3
+
4
+ BASE_PATH=/home/$USER/reComputer
5
+ mkdir -p $BASE_PATH/
6
+ JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
7
+ BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
8
+ if [ -d $JETSON_REPO_PATH ]; then
9
+ echo "jetson-ai-lab existed."
10
+ else
11
+ echo "jetson-ai-lab does not installed. start init..."
12
+ cd $BASE_PATH/
13
+ git clone --depth=1 $BASE_JETSON_LAB_GIT
14
+ cd $JETSON_REPO_PATH
15
+ bash install.sh
16
+ fi
@@ -1,10 +1,10 @@
1
- #!/bin/bash
2
-
3
- BASE_PATH=/home/$USER/reComputer
4
- JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
5
- cd $JETSON_REPO_PATH
6
-
7
- ./run.sh $(./autotag llava) \
8
- python3 -m llava.serve.cli \
9
- --model-path liuhaotian/llava-v1.5-7b \
1
+ #!/bin/bash
2
+
3
+ BASE_PATH=/home/$USER/reComputer
4
+ JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
5
+ cd $JETSON_REPO_PATH
6
+
7
+ ./run.sh $(./autotag llava) \
8
+ python3 -m llava.serve.cli \
9
+ --model-path liuhaotian/llava-v1.5-7b \
10
10
  --image-file /data/images/hoover.jpg
@@ -1,16 +1,16 @@
1
- #!/bin/bash
2
-
3
-
4
- BASE_PATH=/home/$USER/reComputer
5
- mkdir -p $BASE_PATH/
6
- JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
7
- BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
8
- if [ -d $JETSON_REPO_PATH ]; then
9
- echo "jetson-ai-lab existed."
10
- else
11
- echo "jetson-ai-lab does not installed. start init..."
12
- cd $BASE_PATH/
13
- git clone --depth=1 $BASE_JETSON_LAB_GIT
14
- cd $JETSON_REPO_PATH
15
- bash install.sh
16
- fi
1
+ #!/bin/bash
2
+
3
+
4
+ BASE_PATH=/home/$USER/reComputer
5
+ mkdir -p $BASE_PATH/
6
+ JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
7
+ BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
8
+ if [ -d $JETSON_REPO_PATH ]; then
9
+ echo "jetson-ai-lab existed."
10
+ else
11
+ echo "jetson-ai-lab does not installed. start init..."
12
+ cd $BASE_PATH/
13
+ git clone --depth=1 $BASE_JETSON_LAB_GIT
14
+ cd $JETSON_REPO_PATH
15
+ bash install.sh
16
+ fi
@@ -1,11 +1,11 @@
1
- #!/bin/bash
2
-
3
- BASE_PATH=/home/$USER/reComputer
4
- JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
5
- cd $JETSON_REPO_PATH
6
-
7
- ./run.sh $(./autotag local_llm) \
8
- python3 -m local_llm --api=mlc \
9
- --model liuhaotian/llava-v1.6-vicuna-7b \
10
- --max-context-len 768 \
1
+ #!/bin/bash
2
+
3
+ BASE_PATH=/home/$USER/reComputer
4
+ JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
5
+ cd $JETSON_REPO_PATH
6
+
7
+ ./run.sh $(./autotag local_llm) \
8
+ python3 -m local_llm --api=mlc \
9
+ --model liuhaotian/llava-v1.6-vicuna-7b \
10
+ --max-context-len 768 \
11
11
  --max-new-tokens 128
@@ -1,16 +1,16 @@
1
- #!/bin/bash
2
-
3
-
4
- BASE_PATH=/home/$USER/reComputer
5
- mkdir -p $BASE_PATH/
6
- JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
7
- BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
8
- if [ -d $JETSON_REPO_PATH ]; then
9
- echo "jetson-ai-lab existed."
10
- else
11
- echo "jetson-ai-lab does not installed. start init..."
12
- cd $BASE_PATH/
13
- git clone --depth=1 $BASE_JETSON_LAB_GIT
14
- cd $JETSON_REPO_PATH
15
- bash install.sh
16
- fi
1
+ #!/bin/bash
2
+
3
+
4
+ BASE_PATH=/home/$USER/reComputer
5
+ mkdir -p $BASE_PATH/
6
+ JETSON_REPO_PATH="$BASE_PATH/jetson-containers"
7
+ BASE_JETSON_LAB_GIT="https://github.com/dusty-nv/jetson-containers"
8
+ if [ -d $JETSON_REPO_PATH ]; then
9
+ echo "jetson-ai-lab existed."
10
+ else
11
+ echo "jetson-ai-lab does not installed. start init..."
12
+ cd $BASE_PATH/
13
+ git clone --depth=1 $BASE_JETSON_LAB_GIT
14
+ cd $JETSON_REPO_PATH
15
+ bash install.sh
16
+ fi
@@ -1,10 +1,10 @@
1
- # NanoDB
2
-
3
- ## ref
4
-
5
- - <https://www.jetson-ai-lab.com/tutorial_nanodb.html>
6
-
7
- ## access
8
-
9
- - using in machine, try `http://127.0.0.1:7860` in browser.
10
- - using in other pc, make sure you know jetson's IP and try `http://<<Jetson's IP>>:7860` in browser.
1
+ # NanoDB
2
+
3
+ ## ref
4
+
5
+ - <https://www.jetson-ai-lab.com/tutorial_nanodb.html>
6
+
7
+ ## access
8
+
9
+ - using in machine, try `http://127.0.0.1:7860` in browser.
10
+ - using in other pc, make sure you know jetson's IP and try `http://<<Jetson's IP>>:7860` in browser.