Llama3.2:3b randomly outputting "GGGGGGGG" when running under ollama on Jetson Orin Nano Super (JP6.2)

I’m using the Jetson Orin Nano Super to test the effect of running a typical LLM on it.

The ollama and llama3.2:3b was set up under the official tutorial.

The power adaptor is the original one, and I’m using MAXN SUPER power mode without any power supply warning. The bug is after interacting with the llama3.2:3b back-and-forth for several times, the output will suddenly become “GGGGGGGGGGG“ no matter what you typed the last time.

I made a 16GB swap on the SSD, however the phenomenon still persisted even though the RAM hadn’t been filled up.

In spite of all that has been said, the kit itself produced an unpleasant, consistent high-pitch sound (the Coil Whine) when running ollama. Is that normal?

So how can I fix this bug?

Thanks in advance.

*** Please note that this reply is generated by LLM automatically ***

Based on the provided context, I’ll address the issues you’ve encountered with your Jetson Orin Nano Super.

Issue 1: Output becoming “GGGGGGGGGGG” after interacting with llama323b

This issue seems to be related to the power management and thermal settings of your Jetson Orin Nano Super. Although you’re using the MAXN_SUPER power mode, which is designed to provide a balance between performance and power consumption, it’s possible that the system is still experiencing thermal throttling or power limitations.

To troubleshoot this issue, you can try the following:

  1. Monitor system temperatures: Use tools like tmon or jetson_clocks to monitor the system temperatures and ensure they are within the recommended range.
  2. Adjust power settings: Try adjusting the power settings using the nvpmodel tool to see if it improves the situation. You can try setting the power mode to MAXN or MAXP to see if it makes a difference.
  3. Check for software issues: Ensure that the llama323b software is properly configured and optimized for the Jetson Orin Nano Super. You may want to try updating the software or checking for any known issues.

Issue 2: High-pitched sound when running ollama

The high-pitched sound you’re experiencing is likely due to the power supply or the system’s cooling mechanism. Here are some possible causes:

  1. Power supply: The original power adaptor might not be sufficient for the Jetson Orin Nano Super, especially when running demanding workloads like ollama. Try using a more powerful power supply to see if it resolves the issue.
  2. Cooling mechanism: The Jetson Orin Nano Super’s cooling mechanism might be causing the high-pitched sound. You can try adjusting the fan settings or ensuring that the system is properly ventilated.

Issue 3: Swap space and RAM usage

Although you’ve created a 16GB swap on the SSD, it’s possible that the system is still experiencing memory constraints. You can try:

  1. Monitoring RAM usage: Use tools like htop or top to monitor the RAM usage and ensure that it’s not being filled up.
  2. Optimizing software: Ensure that the llama323b software is optimized for the Jetson Orin Nano Super’s hardware and is not consuming excessive memory.

In summary, to fix the issues you’re experiencing, you can try adjusting the power settings, monitoring system temperatures, checking for software issues, and ensuring that the power supply and cooling mechanism are sufficient. Additionally, optimizing the software and monitoring RAM usage can help resolve the issues.

*** Please give the thumbs up if you get answers you like. Or provide feedback to help us improve the answer. ***

Hi,

Could you share more info with us?

Is this reproducible each time you interact with Ollama?
If so, could you share the detailed steps to reproduce this issue?

Is this issue reproducible with other power modes?
Thanks.

Yes, this bug appears each time I interact with llama3.2:3b using the ollama structure.

Moreover, when I open up the Open WebUI to interact, the “GGGGGGG“ hell will immediately appear when I interact with it for the second time (back-and-forth round 2). Next, if I start a new session, the “G“ hell will appear earlier (half way the first round! ) .

This unbearable phenomenon and the coil whine have been going on for three days (exactly till now since I first received my jetson and started using the GPU on it to inference LLMs! )

At first, I wondered that this might be caused by the using up of the RAM, so I swaped a 16GB swap on my SSD, but the bugs persisted.

My power supply source is stable, and I was using a 256GB high-speed SD card as the system card.

To reproduce this, just simply follow the official tutorial on how to run ollama on Jetson Orin Nano Super (JP6.2)

I didn’t use docker. I installed ollama right on my jetson.

Hello,

I’m facing the exact same issue using JetPack 6.2 and running any model in ollama. I have also followed the official tutorial and do not user docker but ollama is directly installed.

This issue appears on a Jeston Orin Nano Super I have just recevied. (I’m using an SSD in my case)

The result could be “GGGGGGGGGGGGG” or “3333333333333” and also see “CONTROL_36CONTROL_36CONTROL_36CONTROL_36CONTROL_36”

Hi

Well I thought that’s a hard fault, so yesterday I returned it to the vendor.

Maybe they will later deliver me a new one to test.

But one more detail to ask, which is the high-pitch coil whine when the GPU is used. Have you ever noticed that? If this phenomenon also occurred on your jetson, the terrible sound might be normal.

Hi,

I have finaly be able to use ollama without any issue on the Jeston using Docker or Jeston containes. Looks the issue is specific to ollama directy installed on the Jestion.

Yes, the fan makes some noise when there is a high load but nothing really wrong for me.

1 Like

No no, I’m not talking about the fan noise, it’s from the capacitor coil.

I guess there’s something wrong with the power, maybe damaged by violent vibrations on the way.

Thank you for your instant report!

I didn’t notice any capacitor coil noise on my Jetson. I hope you will not have the same issue with your new item.

Hhhhhh, thank you for your blessing……

I’ll use docker later.

Let the record show that I am having very similar issues on my Jetson Orin AGX 64Gb Developer kit.

I will attempt to use the docker container instead and report back here.

UGHHH it works with the jetson-containers run --name ollama $(autotag ollama) command. Here is my problem with this though.

It is using ollama version 0.6.8 which is OLD and doesnt have access to some of the newer models I want to run.

I think my work around will be this:

  • Run the jetson-container ollama 0.6.8 on the regular 11434 port
    • I will host my smaller models like llama3.2 and gemma3 on here.
  • Run my natively installed ollama on port 9000.
    • I will host my newer and larger models like gpt-oss:20b and llama3.2-vision on this one.

If someone has a better idea, I am all ears.

In my case the output is 3333333 of a repeated < unused31>.

Good to hear that the docker method seems to solve issues. Is there a clear step by step instruction for this?

I do hope people are working on fixing the issue, because running ollama natively should not be any big task as weaker hardware copes just fine (al be it with slower performance)

I have followed the instructions on https://www.youtube.com/watch?v=R0PjKr4d-gU to get a docker running ollama, but I still get the < unused31 > repeated output for a longer prompt (that works on weaker hardware)

The other problem is that when the llms do run they make weird statements that hint at them ‘forgetting’ parts of what was input or not applying things they ‘know’. And again, compared to the same models running on another system, these problem occur on only on the jetson, and not in similar circumstances on the other (weaker) hardware like a Pi5.

And finally, installing a slightly bigger model that is stated to work on the hardware it gives errors both natively and in a docker

update, I decided to try a clean install of only the official ubuntu image and only the steps from the working instructions.

Unfortunately, it still gives the same error although the llama3.1 I also tried just started to output rather random words and pars of words it tried to organise into sentences

Is it solved? Same issue happend on my Orin. I tested qwen2.5:7b qwen3:8b llama3:8b deepseek-r1-1.5B, they all have this issue with a long prompt. The only model works is gemma3:1b for some unknown reason

No, this has NOT been solved (and keeps happening with multiple models, both large and small). Thanks for keeping the thread alive.

I have not seen a solution yet and multiple models from small ones to big ones eventually stop working. On my current install I have not set any swap file, so caching issues there are not at fault.

Docker does not solve it and uses extra memory, which might even complicate things.

Hope someone can fix this soon. I tried the docker solution, it magically made qwen3:8b working but didn’t help with other models.

1 Like