8,629
edits
Paradox-01 (talk | contribs) mNo edit summary |
Paradox-01 (talk | contribs) mNo edit summary |
||
| Line 564: | Line 564: | ||
** The actual data-holding model, including its parameters and weights. Most often, this is a Large Language Model (LLM) or a Large Multimodal Model (LMM). The learned data consist of statistical patterns about text, images, or other media. Very most of the original raw data cannot be reconstructed from the model. | ** The actual data-holding model, including its parameters and weights. Most often, this is a Large Language Model (LLM) or a Large Multimodal Model (LMM). The learned data consist of statistical patterns about text, images, or other media. Very most of the original raw data cannot be reconstructed from the model. | ||
** Reinforcement learning from human feedback (RLHF) and (its successor RLAIF) can be named as another important feature that added a reward model for higher quality and alignment. | ** Reinforcement learning from human feedback (RLHF) and (its successor RLAIF) can be named as another important feature that added a reward model for higher quality and alignment. | ||
** Other features like COT, MoE, the use of external tools | ** Other features or milestones like COT, MoE, context expansion and the use of external tools via MCP to compensate own shortcomings are better described as incremental improvements in the evolution of GenAI.<!-- | ||
* GPT = Generative pre-trained transformers (Large Language Model with the actual "learning" part de facto outsourced to humans: Reinforcement learning from human feedback (RLHF), in best case GPTs have a ''transplanted base intelligence'' but they lack the important feature to really learn for themselves. Low quality "synthetic data" can even worsen the models. | * GPT = Generative pre-trained transformers (Large Language Model with the actual "learning" part de facto outsourced to humans: Reinforcement learning from human feedback (RLHF), in best case GPTs have a ''transplanted base intelligence'' but they lack the important feature to really learn for themselves. Low quality "synthetic data" can even worsen the models. | ||
:: After the obvious slowdown in advancement through scaling, chain of thought (COT) was introduced. It is also known under the marketing term "reasoning [AI]". (See: Gemini 2.0 Flash Thinking and ChatGPT o3, the later one especially [https://techxplore.com/news/2024-12-ai-human-general-intelligence.html trained to cheat … erm ... score high in the ARC-AGI test].) | :: After the obvious slowdown in advancement through scaling, chain of thought (COT) was introduced. It is also known under the marketing term "reasoning [AI]". (See: Gemini 2.0 Flash Thinking and ChatGPT o3, the later one especially [https://techxplore.com/news/2024-12-ai-human-general-intelligence.html trained to cheat … erm ... score high in the ARC-AGI test].) | ||
edits