<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Run LLaMA-Factory on edge devices Archives - OpenZeka EN Blog</title>
	<atom:link href="https://blog-en.openzeka.com/tag/run-llama-factory-on-edge-devices/feed/" rel="self" type="application/rss+xml" />
	<link>https://blog-en.openzeka.com/tag/run-llama-factory-on-edge-devices/</link>
	<description>NVIDIA Jetson Developer Kits &#38;Edge Devices</description>
	<lastBuildDate>Fri, 27 Mar 2026 13:27:00 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>
	<item>
		<title>Jetson Generative AI – LLaMA Factory</title>
		<link>https://blog-en.openzeka.com/jetson-generative-ai-llama-factory/</link>
		
		<dc:creator><![CDATA[Enhar]]></dc:creator>
		<pubDate>Thu, 31 Jul 2025 05:09:53 +0000</pubDate>
				<category><![CDATA[Generative AI]]></category>
		<category><![CDATA[Fine-tune LLaMA models on Jetson]]></category>
		<category><![CDATA[Jetson LLaMA-Factory tutorial]]></category>
		<category><![CDATA[LLaMA-Factory Jetson setup]]></category>
		<category><![CDATA[Local LLM fine-tuning Jetson]]></category>
		<category><![CDATA[Run LLaMA-Factory on edge devices]]></category>
		<guid isPermaLink="false">https://blog.aetherix.com/?p=1055</guid>

					<description><![CDATA[<p>LLaMA Factory provides a unified framework for fine-tu ... Continue Reading→</p>
<p>The post <a href="https://blog-en.openzeka.com/jetson-generative-ai-llama-factory/">Jetson Generative AI – LLaMA Factory</a> appeared first on <a href="https://blog-en.openzeka.com">OpenZeka EN Blog</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="fusion-fullwidth fullwidth-box fusion-builder-row-1 fusion-flex-container has-pattern-background has-mask-background nonhundred-percent-fullwidth non-hundred-percent-height-scrolling" style="--awb-border-radius-top-left:0px;--awb-border-radius-top-right:0px;--awb-border-radius-bottom-right:0px;--awb-border-radius-bottom-left:0px;--awb-flex-wrap:wrap;" ><div class="fusion-builder-row fusion-row fusion-flex-align-items-flex-start fusion-flex-content-wrap" style="max-width:1331.2px;margin-left: calc(-4% / 2 );margin-right: calc(-4% / 2 );"><div class="fusion-layout-column fusion_builder_column fusion-builder-column-0 fusion_builder_column_1_1 1_1 fusion-flex-column" style="--awb-bg-size:cover;--awb-width-large:100%;--awb-margin-top-large:0px;--awb-spacing-right-large:1.92%;--awb-margin-bottom-large:20px;--awb-spacing-left-large:1.92%;--awb-width-medium:100%;--awb-order-medium:0;--awb-spacing-right-medium:1.92%;--awb-spacing-left-medium:1.92%;--awb-width-small:100%;--awb-order-small:0;--awb-spacing-right-small:1.92%;--awb-spacing-left-small:1.92%;"><div class="fusion-column-wrapper fusion-column-has-shadow fusion-flex-justify-content-flex-start fusion-content-layout-column"><div class="fusion-text fusion-text-1"><p>LLaMA Factory provides a unified framework for fine-tuning large language models with an intuitive web interface. This powerful tool brings professional model training capabilities to Jetson devices, enabling you to customize LLMs for your specific use cases with optimized performance for edge deployment.</p>
<p>In this article, you&#8217;ll learn how to run LLaMA Factory on Jetson Orin for <strong>efficient LLM fine-tuning and deployment.</strong></p>
</div><div class="fusion-image-element " style="--awb-caption-title-font-family:var(--h2_typography-font-family);--awb-caption-title-font-weight:var(--h2_typography-font-weight);--awb-caption-title-font-style:var(--h2_typography-font-style);--awb-caption-title-size:var(--h2_typography-font-size);--awb-caption-title-transform:var(--h2_typography-text-transform);--awb-caption-title-line-height:var(--h2_typography-line-height);--awb-caption-title-letter-spacing:var(--h2_typography-letter-spacing);"><span class=" fusion-imageframe imageframe-none imageframe-1 hover-type-none"><img fetchpriority="high" decoding="async" width="1024" height="675" title="llama_factory_interface" src="https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_interface-1024x675.webp" alt class="img-responsive wp-image-1059" srcset="https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_interface-200x132.webp 200w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_interface-300x198.webp 300w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_interface-400x264.webp 400w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_interface-600x395.webp 600w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_interface-768x506.webp 768w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_interface-800x527.webp 800w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_interface-1024x675.webp 1024w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_interface-1200x791.webp 1200w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_interface-1536x1012.webp 1536w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_interface.webp 1554w" sizes="(max-width: 640px) 100vw, 1024px" /></span></div><div class="fusion-title title fusion-title-1 fusion-sep-none fusion-title-text fusion-title-size-four" style="--awb-margin-bottom:-20px;"><h4 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>Features</h4></h4></div><div class="fusion-text fusion-text-2"><ul>
<li>Support for multiple LLM architectures including LLaMA, Qwen, ChatGLM, and more</li>
<li>Multiple training stages: Supervised Fine-Tuning, Reward</li>
<li>Modeling, PPO, DPO, KTO, Pre-Training</li>
<li>Three fine-tuning methods: full, freeze, and lora</li>
<li>Gradio-based web UI with Train, Evaluate &amp; Predict, Chat, and</li>
<li>Export tabs</li>
<li>Built-in dataset support with preview functionality</li>
<li>Integrated chat interface for testing models</li>
<li>Real-time training loss visualization</li>
<li>Advanced configurations for quantization, LoRA, RLHF, and more</li>
<li>Model evaluation with customizable generation parameters</li>
</ul>
</div><div class="fusion-title title fusion-title-2 fusion-sep-none fusion-title-text fusion-title-size-four" style="--awb-margin-bottom:-20px;"><h4 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>Requirements</h4></h4></div>
<div class="table-1">
<p>&nbsp;</p>
<table width="100%">
<thead>
<tr>
<th align="left">
<div>
<div>Hardware / Software</div>
</div>
</th>
<th align="left">
<div>
<div>Notes</div>
</div>
</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">
<div>
<div><strong>Jetson AGX Orin with ≥ 32 GB RAM</strong></div>
</div>
</td>
<td align="left">64GB recommended for larger models</td>
</tr>
<tr>
<td align="left">
<div>
<div><strong>JetPack 6.0+</strong></div>
</div>
</td>
<td align="left">For CUDA 12.x support</td>
</tr>
<tr>
<td align="left">
<div>
<div><strong>NVMe SSD</strong></div>
</div>
</td>
<td align="left">Essential for model storage and caching</td>
</tr>
<tr>
<td align="left">
<div>
<div><strong>Hugging Face token</strong></div>
</div>
</td>
<td align="left">Required for accessing gated models</td>
</tr>
<tr>
<td align="left">
<div>
<div><strong>~50 GB free storage</strong></div>
</div>
</td>
<td align="left">For models and training checkpoints</td>
</tr>
</tbody>
</table>
</div>
<div class="fusion-title title fusion-title-3 fusion-sep-none fusion-title-text fusion-title-size-four"><h4 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>Step-by-Step Setup</h4></h4></div><div class="fusion-title title fusion-title-4 fusion-sep-none fusion-title-text fusion-title-size-four"><h4 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>1. Create necessary directories</h4></h4></div><style type="text/css" scopped="scopped">.fusion-syntax-highlighter-1 > .CodeMirror, .fusion-syntax-highlighter-1 > .CodeMirror .CodeMirror-gutters {background-color:#2d3748;}</style><div class="fusion-syntax-highlighter-container fusion-syntax-highlighter-1 fusion-syntax-highlighter-theme-dark" style="opacity:0;margin-top:0px;margin-right:0px;margin-bottom:0px;margin-left:0px;font-size:14px;border-width:1px;border-style:solid;border-color:rgba(242,243,245,0);"><div class="syntax-highlighter-copy-code"><span class="syntax-highlighter-copy-code-title" data-id="fusion_syntax_highlighter_1" style="font-size:14px;">Copy to Clipboard</span></div><label for="fusion_syntax_highlighter_1" class="screen-reader-text">Syntax Highlighter</label><textarea class="fusion-syntax-highlighter-textarea" id="fusion_syntax_highlighter_1" data-readOnly="nocursor" data-lineNumbers="" data-lineWrapping="" data-theme="oceanic-next" data-mode="text/x-sh">sudo mkdir -p /mnt/nvme/cache/llama-factory/{cache,config,data,saves}
sudo chown -R $USER:$USER /mnt/nvme/cache/llama-factory</textarea></div><div class="fusion-title title fusion-title-5 fusion-sep-none fusion-title-text fusion-title-size-four"><h4 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>2. Set your Hugging Face token</h4></h4></div><div class="fusion-text fusion-text-3"><p>Replace your_hf_token_here with your actual token from <strong style="color: #0fbc00;">https://huggingface.co/settings/tokens</strong><span style="color: #0fbc00;"> .</span></p>
</div><style type="text/css" scopped="scopped">.fusion-syntax-highlighter-2 > .CodeMirror, .fusion-syntax-highlighter-2 > .CodeMirror .CodeMirror-gutters {background-color:#2d3748;}</style><div class="fusion-syntax-highlighter-container fusion-syntax-highlighter-2 fusion-syntax-highlighter-theme-dark" style="opacity:0;margin-top:0px;margin-right:0px;margin-bottom:0px;margin-left:0px;font-size:14px;border-width:1px;border-style:solid;border-color:rgba(242,243,245,0);"><div class="syntax-highlighter-copy-code"><span class="syntax-highlighter-copy-code-title" data-id="fusion_syntax_highlighter_2" style="font-size:14px;">Copy to Clipboard</span></div><label for="fusion_syntax_highlighter_2" class="screen-reader-text">Syntax Highlighter</label><textarea class="fusion-syntax-highlighter-textarea" id="fusion_syntax_highlighter_2" data-readOnly="nocursor" data-lineNumbers="" data-lineWrapping="" data-theme="oceanic-next" data-mode="text/x-sh">export HF_TOKEN=your_hf_token_here</textarea></div><div class="fusion-title title fusion-title-6 fusion-sep-none fusion-title-text fusion-title-size-four"><h4 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>3. Launch LLaMA Factory</h4></h4></div><style type="text/css" scopped="scopped">.fusion-syntax-highlighter-3 > .CodeMirror, .fusion-syntax-highlighter-3 > .CodeMirror .CodeMirror-gutters {background-color:#2d3748;}</style><div class="fusion-syntax-highlighter-container fusion-syntax-highlighter-3 fusion-syntax-highlighter-theme-dark" style="opacity:0;margin-top:0px;margin-right:0px;margin-bottom:0px;margin-left:0px;font-size:14px;border-width:1px;border-style:solid;border-color:rgba(242,243,245,0);"><div class="syntax-highlighter-copy-code"><span class="syntax-highlighter-copy-code-title" data-id="fusion_syntax_highlighter_3" style="font-size:14px;">Copy to Clipboard</span></div><label for="fusion_syntax_highlighter_3" class="screen-reader-text">Syntax Highlighter</label><textarea class="fusion-syntax-highlighter-textarea" id="fusion_syntax_highlighter_3" data-readOnly="nocursor" data-lineNumbers="" data-lineWrapping="" data-theme="oceanic-next" data-mode="text/x-sh">docker run -it --rm \
  --name=llama-factory \
  -v /mnt/nvme/cache/llama-factory/cache:/data/llama-factory/cache \
  -v /mnt/nvme/cache/llama-factory/config:/data/llama-factory/config \
  -v /mnt/nvme/cache/llama-factory/data:/data/llama-factory/data \
  -v /mnt/nvme/cache/llama-factory/saves:/data/llama-factory/saves \
  -e GRADIO_SERVER_PORT=7860 \
  -e API_PORT=9000 \
  -p 7860:7860 \
  -p 9000:9000 \
  --runtime=nvidia \
  -e DOCKER_PULL=always --pull always \
  -e HF_TOKEN=$HF_TOKEN \
  -e HF_HUB_CACHE=/root/.cache/huggingface \
  -v /mnt/nvme/cache:/root/.cache \
  dustynv/llama-factory:r36.3.0</textarea></div><div class="fusion-title title fusion-title-7 fusion-sep-none fusion-title-text fusion-title-size-one"><h1 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>4. Access the Web UI</h4></h1></div><div class="fusion-text fusion-text-4"><p>Once the container starts, you&#8217;ll see:</p>
<blockquote>
<p>Running on local URL: http://0.0.0.0:7860</p>
</blockquote>
<p><strong>Local access:</strong> Open <strong>http://localhost:7860</strong> in your browser<br />
<strong>Remote access:</strong> Use<strong> http://&lt;jetson-ip&gt;:7860</strong></p>
</div><div class="fusion-image-element " style="--awb-caption-title-font-family:var(--h2_typography-font-family);--awb-caption-title-font-weight:var(--h2_typography-font-weight);--awb-caption-title-font-style:var(--h2_typography-font-style);--awb-caption-title-size:var(--h2_typography-font-size);--awb-caption-title-transform:var(--h2_typography-text-transform);--awb-caption-title-line-height:var(--h2_typography-line-height);--awb-caption-title-letter-spacing:var(--h2_typography-letter-spacing);"><span class=" fusion-imageframe imageframe-none imageframe-2 hover-type-none"><img decoding="async" width="1024" height="769" title="llama_factory_training" src="https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_training-1024x769.webp" alt class="img-responsive wp-image-1060" srcset="https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_training-200x150.webp 200w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_training-300x225.webp 300w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_training-400x300.webp 400w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_training-600x451.webp 600w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_training-768x577.webp 768w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_training-800x601.webp 800w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_training-1024x769.webp 1024w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_training-1200x901.webp 1200w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_training-1536x1153.webp 1536w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_training.webp 1550w" sizes="(max-width: 640px) 100vw, 1024px" /></span></div><div class="fusion-title title fusion-title-8 fusion-sep-none fusion-title-text fusion-title-size-four"><h4 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>5. Configure your fine-tuning</h4></h4></div><div class="fusion-text fusion-text-5"><p>In the <strong>Train</strong> tab:</p>
<p><strong>    1.Select Training Stage:</strong></p>
<ul>
<li>Supervised Fine-Tuning (most common)</li>
<li>Reward Modeling</li>
<li>PPO (Proximal Policy Optimization)</li>
<li>DPO (Direct Preference Optimization)</li>
<li>KTO</li>
<li>Pre-Training</li>
</ul>
<p><strong>    2.Choose Finetuning Method:</strong></p>
<ul>
<li>lora &#8211; Low-rank adaptation, best for memory efficiency</li>
<li>freeze &#8211; Freezes base model layers</li>
<li>full &#8211; Full parameter fine-tuning</li>
</ul>
<p><strong>    </strong><strong>3.Configure Data:</strong></p>
<ul>
<li>Data directory: /opt/LLaMA-Factory/data</li>
<li>Select dataset from dropdown</li>
<li>Use &#8220;Preview dataset&#8221; to verify data format</li>
</ul>
<p><strong>    4.Set Training Parameters:</strong></p>
<ul>
<li>Cutoff length: 1024 (max tokens in input sequence)</li>
<li>Max samples: 100000</li>
<li>Batch size: 2</li>
<li>Learning rate: 5e-5 (in Advanced configurations)</li>
<li>Epochs: 3.0 (in Advanced configurations)</li>
</ul>
<p><strong>    5.Advanced Configurations (expandable sections):</strong></p>
<ul>
<li>Quantization bit (none/bitsandbytes)</li>
<li>Extra configurations</li>
<li>Freeze tuning configurations</li>
<li>LoRA configurations</li>
<li>RLHF configurations</li>
<li>GaLore configurations</li>
<li>BAdam configurations</li>
</ul>
<p><strong>    6.Start Training:</strong></p>
<ul>
<li>Click &#8220;Preview command&#8221; to verify settings</li>
<li>Click &#8220;Start&#8221; to begin training</li>
<li>Monitor real-time loss graph</li>
</ul>
</div><div class="fusion-title title fusion-title-9 fusion-sep-none fusion-title-text fusion-title-size-one"><h1 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>6. Evaluate your model</h4></h1></div><div class="fusion-text fusion-text-6"><p>Switch to the Evaluate &amp; Predict tab to assess model performance:</p>
<p><strong>1.Configure evaluation settings:</strong></p>
<ul>
<li>Data directory and dataset (same as training)</li>
<li>Cutoff length: 1024</li>
<li>Max samples: 100000</li>
<li>Batch size: 2</li>
</ul>
<p><strong>2.Set generation parameters:</strong></p>
<ul>
<li>Maximum new tokens: 512</li>
<li>Top-p: 0.7</li>
<li>Temperature: 0.95</li>
</ul>
<p><strong>3.Run evaluation:</strong></p>
<ul>
<li>Enable &#8220;Save predictions&#8221; to store results</li>
<li>Click &#8220;Start&#8221; to begin evaluation</li>
<li>Results saved to timestamped output directory</li>
</ul>
</div><div class="fusion-image-element " style="--awb-caption-title-font-family:var(--h2_typography-font-family);--awb-caption-title-font-weight:var(--h2_typography-font-weight);--awb-caption-title-font-style:var(--h2_typography-font-style);--awb-caption-title-size:var(--h2_typography-font-size);--awb-caption-title-transform:var(--h2_typography-text-transform);--awb-caption-title-line-height:var(--h2_typography-line-height);--awb-caption-title-letter-spacing:var(--h2_typography-letter-spacing);"><span class=" fusion-imageframe imageframe-none imageframe-3 hover-type-none"><img decoding="async" width="1024" height="358" title="llama_factory_evaluate" src="https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_evaluate-1024x358.webp" alt class="img-responsive wp-image-1058" srcset="https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_evaluate-200x70.webp 200w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_evaluate-300x105.webp 300w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_evaluate-400x140.webp 400w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_evaluate-600x210.webp 600w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_evaluate-768x269.webp 768w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_evaluate-800x280.webp 800w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_evaluate-1024x358.webp 1024w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_evaluate-1200x420.webp 1200w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/llama_factory_evaluate.webp 1532w" sizes="(max-width: 640px) 100vw, 1024px" /></span></div><div class="fusion-title title fusion-title-10 fusion-sep-none fusion-title-text fusion-title-size-four"><h4 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>7. Test your model interactively</h4></h4></div><div class="fusion-text fusion-text-7"><p>Navigate to the Chat tab:</p>
<p><strong>1.Load your model:</strong></p>
<ul>
<li>Click &#8220;Load model&#8221; button</li>
<li>Select inference engine: huggingface</li>
<li>Choose inference data type: auto</li>
</ul>
<p><strong>2.Chat with your model:</strong></p>
<ul>
<li>Type messages in the chat interface</li>
<li>Model responds in real-time</li>
<li>Test both base and fine-tuned versions</li>
</ul>
<p><strong>3.Unload model when switching between models</strong></p>
</div><div class="fusion-image-element " style="--awb-caption-title-font-family:var(--h2_typography-font-family);--awb-caption-title-font-weight:var(--h2_typography-font-weight);--awb-caption-title-font-style:var(--h2_typography-font-style);--awb-caption-title-size:var(--h2_typography-font-size);--awb-caption-title-transform:var(--h2_typography-text-transform);--awb-caption-title-line-height:var(--h2_typography-line-height);--awb-caption-title-letter-spacing:var(--h2_typography-letter-spacing);"><span class=" fusion-imageframe imageframe-none imageframe-4 hover-type-none"><img decoding="async" width="1024" height="539" title="chatinterface" src="https://blog-en.openzeka.com/wp-content/uploads/2025/07/chatinterface-1024x539.webp" alt class="img-responsive wp-image-1057" srcset="https://blog-en.openzeka.com/wp-content/uploads/2025/07/chatinterface-200x105.webp 200w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/chatinterface-300x158.webp 300w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/chatinterface-400x210.webp 400w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/chatinterface-600x316.webp 600w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/chatinterface-768x404.webp 768w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/chatinterface-800x421.webp 800w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/chatinterface-1024x539.webp 1024w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/chatinterface-1200x631.webp 1200w, https://blog-en.openzeka.com/wp-content/uploads/2025/07/chatinterface-1536x808.webp 1536w" sizes="(max-width: 640px) 100vw, 1024px" /></span></div><div class="fusion-title title fusion-title-11 fusion-sep-none fusion-title-text fusion-title-size-four"><h4 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>8. Export your model</h4></h4></div><div class="fusion-text fusion-text-8"><p>Use the <strong>Export</strong> tab to save your fine-tuned model in various formats for deployment. This allows you to use your model outside of <strong>LLaMA Factory</strong> in production environments.</p>
</div><div class="fusion-title title fusion-title-12 fusion-sep-none fusion-title-text fusion-title-size-four"><h4 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>Training Method Guide</h4></h4></div>
<div class="table-1">
<table width="100%">
<thead>
<tr>
<th align="left">Method</th>
<th align="left">Memory Usage</th>
<th align="left">Training Speed</th>
<th align="left">Use Case</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left"><strong>LoRA</strong></td>
<td align="left">Low</td>
<td align="left">Fast</td>
<td align="left">Recommended for most Jetson deployments</td>
</tr>
<tr>
<td align="left"><strong>Freeze</strong></td>
<td align="left">Medium</td>
<td align="left">Medium</td>
<td align="left">When you need to preserve base model behavior</td>
</tr>
<tr>
<td align="left"><strong>Full</strong></td>
<td align="left">High</td>
<td align="left">Slow</td>
<td align="left">Small models only (≤1.5B parameters)</td>
</tr>
</tbody>
</table>
</div>
<div class="fusion-title title fusion-title-13 fusion-sep-none fusion-title-text fusion-title-size-four" style="--awb-margin-bottom:-20px;"><h4 class="fusion-title-heading title-heading-left" style="margin:0;"><h4>Troubleshooting</h4></h4></div>
<div class="table-1">
<p>&nbsp;</p>
<table width="100%">
<thead>
<tr>
<th align="left">Issue</th>
<th align="left">Fix</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left"><strong>Out of memory during training</strong></td>
<td align="left">Reduce batch size to 1-2, use LoRA method, or use smaller model</td>
</tr>
<tr>
<td align="left"><strong>Slow model download</strong></td>
<td align="left">Models are cached in <code>/mnt/nvme/cache/huggingface</code>, be patient on first run</td>
</tr>
<tr>
<td align="left"><strong>Connection refused</strong></td>
<td align="left">Ensure port 7860 is not blocked by firewall</td>
</tr>
<tr>
<td align="left"><strong>Training won&#8217;t start</strong></td>
<td align="left">Check dataset format matches the selected template</td>
</tr>
<tr>
<td align="left"><strong>GPU not utilized</strong></td>
<td align="left">Verify with <code>tegrastats</code> and ensure <code>--runtime nvidia</code> is set</td>
</tr>
</tbody>
</table>
</div>
<div class="fusion-text fusion-text-9"><p>&nbsp;</p>
<p><strong>For more information</strong> about LLaMA Factory features and supported models, visit the<a href="https://github.com/hiyouga/LLaMA-Factory"><strong style="color: #0fbc00;"> LLaMA Factory repository.</strong></a></p>
</div></div></div></div></div>
<p>The post <a href="https://blog-en.openzeka.com/jetson-generative-ai-llama-factory/">Jetson Generative AI – LLaMA Factory</a> appeared first on <a href="https://blog-en.openzeka.com">OpenZeka EN Blog</a>.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
