Almost the same could be told for Unity or any other complex engine/tool

Main content

TLDR

I (and most experienced people on the Unreal Source Discord server) recommend not using any LLMs or AI assistants, especially if you’re a beginner.

  • LLMs are bad at C++ and even more at Unreal Engine methods and systems.
  • LLMs are wrong more often than not.
  • LLMs answer with complete confidence even when wrong.
  • If you’re new you likely don’t known enough to know when answers are good or wrong.

Please read before going further

I will mostly mention ChatGPT to keep the paragraphs short, but please keep in mind that I am talking about any LLMs.

Regarding UE, ChatGPT isn’t recommended, but if you really want to use it for starters, it isn’t recommended to use it once you know the basics of UE.

Why ?
Well any human can read the source code of UE, where there is everything you need, but AI models can’t be trained on it because the source code is “private” (you need to create an Epic and GitHub account, link them and agree to the Unreal Engine EULA).
This means that ChatGPT trained itself on publicly available sources, which are only the Official Docs, forums such as the official one and Reddit, blogs, and maybe YouTube video transcripts. It can sometimes have some source code snippets that users copied from the source, but that’s it.

The biggest issue is that forums posts holds the vast majority of data, and the most used forums publicly available are the official one and Reddit. This would mean that a big % of the answers ChatGPT gives you when asking a UE question comes from posts and comments user wrote on those forums.
And this leads to an issue, forums posts are written by a lot of different kind of people, people asking, people answering right and people saying bad stuff. ChatGPT can hardly tell what is a good or bad answer, also, a lot of posts are old (between 2015-2020), which means some of the good answers can be irrelevant now because the methods are deprecated or changed.

A good example that shows that ChatGPT gives bad answers is this simple question.
Also, people that read a lot of forums posts on any engine can confirm that maybe 20% to 40% of answers are not correct or ideal. This helps you imagine how wrong any AI trained on that data can be.

Even locally trained AI such as Github Copilot or Jetbrain’s AI Assistant are not “good enough” in most cases for experienced people in the engine. They are usually good for single line completion (and in some specific cases good for a few more lines).

Conclusion

Try to not use ChatGPT or any other LLMs AI on complexes engines/tools such as Unreal Engine, some can be less worse than ChatGPT like Perplexity because it gives you the used sources. So you can at least check the context of the answer.

But please learn by yourself! There is no better way to learn and improve, it’s okay to be slow or to not do the correct thing the first times.
It is very important to learn doing research’s on the internet (or elsewhere) by yourself to find answers, this makes you better at finding how to express your thoughts, find keywords and improve your criticism.

Joviex

“Mastery of a foundational skill precedes effective use of advanced tools.”

And there is a ton of friendly humans that can answer your questions with more knowledge and correctness that any LLMs could ever have (for now), find them in the correct forums or communities (shout-out to the Unreal Source Discord and Unity Official Discord).

More on the subject

Disclaimer! I am not an expert in this domain, most of my insights comes from people with experience in the mentioned domains, please be cautious when reading the listed links. Before writing this I mostly used trustworthy articles and my personal (and others) experiences.

LM tech limitations

Thanks to Blue Man from the Unreal Source Discord for this long explanation that he gave me after I asked about the real tech limitation that the industry is currently facing, affecting UE and other complex tools and codebases

Blue Man response

Tldr; is that current AI leaders are pretending there are no limits or flaws because if they tell the truth they would lose investors

The truth is that the attention mechanism that is at the core of every llm can’t scale forever

Attention is basically a fancy dot product, a query (latest token) has to attend to every past token (keys), basically doing a dot product which gives you an attention map of how much each token contributes to the output (and then the result is written to the “residual stream”, if anyone is curious what a residual stream is it’s kinda a fun one to explain so..)

But since it’s basically a dot product each token contributes something, if you have too many tokens competing for attention the signal of actually important tokens gets lost in the noise

That’s why pretty much every single LLM today sucks at anything long context related and that’s not gonna change

But that’s part of the problem Current trend of reinforcement learning to achieve reasoning is also reaching the limit, you can see that by looking at all frontier models today, they all converge at about the exact same benchmark numbers, which is a sign the scaling is reaching its limit

RL is the holy grail of ML, you can do with it what you can’t with a hand crafted dataset but you have a verifiability problem

To train a model with RL the problem needs to be easily verifiable to provide a clear reward signal to the model

And it’s kinda hard to make a verifier for an open ended problem

For programming unit tests are used, but that’s why you now have models that attempt to rewrite half of the codebase when it can’t solve a problem, the reward signal thought it as long as the result passes nothing else matters

So new models today mostly rely on benchmark maxxing (training specifically to beat benchmarks) which doesn’t translate to real world performance

So the llm space right now is kinda at a standstill when it comes to top model progress. Thu the move to agents, where the llm isn’t improving much but the scaffolding around it is trying to squeeze more out of the llm

What’s progressing is small models, you can now run a 32B local model on 1 GPU that will be almost on par with best of the best

Analysis published not long before his message that matches his observations and insight: [How far can reasoning models scale?](https://epoch.ai/gradient-updates/how-far-can-reasoning-models-scale)