Large language model: Difference between revisions

Content deleted Content added
Yoderj (talk | contribs)
Add "decoder-only"
Clyrmze (talk | contribs)
Added a new row under the LLM models list for the newly released Google Gemma.
Line 462:
|Unknown|| {{no|Proprietary}}
| Multimodal model, based on a Mixture-of-Experts (MoE) architecture. Context window increased to 1 million tokens, though only 128k will be available for developers.<ref>{{cite web |title=Our next-generation model: Gemini 1.5 |url=https://blog.google/technology/ai/google-gemini-next-generation-model-february-2024/#context-window |website=Google |access-date=16 February 2024 |language=en-us |date=15 February 2024 |quote=This means 1.5 Pro can process vast amounts of information in one go — including 1 hour of video, 11 hours of audio, codebases with over 30,000 lines of code or over 700,000 words. In our research, we’ve also successfully tested up to 10 million tokens.}}</ref>
|-
| Gemma || {{dts|February 2024}} || [[Google DeepMind]] || 2B and 7B || 6T tokens || Unknown || {{yes|Apache 2.0}}<ref name="gemma">{{cite web|url=https://github.com/google-deepmind/gemma/blob/main/LICENSE|title=Gemma License|via=GitHub}}</ref>||
|}
 
== See also ==
* [[Foundation models]]