Skip Navigation

GPT-4's details are leaked.

cross-posted from: https://lemmy.intai.tech/post/72919

Parameters count:

GPT-4 is more than 10x the size of GPT-3. We believe it has a total of ~1.8 trillion parameters across 120 layers. Mixture Of Experts - Confirmed.

OpenAI was able to keep costs reasonable by utilizing a mixture of experts (MoE) model. They utilizes 16 experts within their model, each is about ~111B parameters for MLP. 2 of these experts are routed to per forward pass.

Related Article: https://lemmy.intai.tech/post/72922

TechNews @radiation.party

GPT-4 Details Leaked

10 0
AI Companions @lemmy.world

GPT-4's details are leaked.

-1 0
Natural Language Programming | Prompting (chatGPT) @lemmy.intai.tech

GPT-4's details are leaked.

1 0
26 comments
  • "Half of those additions are censors and more creative ways to say 'sorry, I can't do that for you Jim.'" Lol, I'm just kidding, 1.8t parameters is incredible.

    I just really hope that it's not as censored as it currently is. ;_;

  • The interesting part to me:

    The missing dataset it a custom dataset of college textbooks collected by hand for as much courses as possible.

    This is very easy to convert to txt file and than with self-instruct into instruction form. This creates the "illusion" that GPT-4 "is smart" no matter who use it.

    Computer scientist? sure! it can help you with your questions about P!=NP Philosophy major? It can totally talk to you about epistemology.

    Don't you see? It was trained on the textbooks. It is so obvious.

    This could explain some (but not all) of the 'magic' I have seen with GPT4 vs GPT3.

    If you put a bunch of textbooks into Google, it still couldn't help me build a video game engine

26 comments