|
| 1 | +created_by: barny |
| 2 | +version: 3 |
| 3 | +domain: large-language-model |
| 4 | +document_outline: Knowledge contribution about the IBM Granite Model |
| 5 | +seed_examples: |
| 6 | + - context: >- |
| 7 | + IBM Granite is a series of decoder-only AI foundation models created by |
| 8 | + IBM.[3] It was announced on September 7, 2023,[4][5] and an initial paper |
| 9 | + was published 4 days later.[6] Initially intended for use in the IBM's |
| 10 | + cloud-based data and generative AI platform Watsonx along with other |
| 11 | + models,[7] IBM opened the source code of some code models.[8][9] Granite |
| 12 | + models are trained on datasets curated from Internet, academic |
| 13 | + publishings, code datasets, legal and finance documents |
| 14 | + questions_and_answers: |
| 15 | + - question: What is IBM Granite? |
| 16 | + answer: >- |
| 17 | + IBM Granite is a series of decoder-only AI foundation models created |
| 18 | + by IBM. |
| 19 | + - question: When as was IBM Granite announced? |
| 20 | + answer: September 7, 2023 |
| 21 | + - question: What's a series of IBM decoder-only AI foundation models? |
| 22 | + answer: IBM Granite |
| 23 | + - context: >- |
| 24 | + Granite's first foundation models were Granite.13b.instruct and |
| 25 | + Granite.13b.chat. The "13b" in their name comes from 13 billion, the |
| 26 | + amount of parameters they have as models, lesser than most of the larger |
| 27 | + models of the time. Later models vary from 3 to 34 billion parameters. |
| 28 | + questions_and_answers: |
| 29 | + - question: what were the first foundation models in IBm Granite? |
| 30 | + answer: Granite.13b.instruct and Granite.13b.chat |
| 31 | + - question: >- |
| 32 | + How many model parameters do Granite.13b.instruct and Granite.13b.chat |
| 33 | + have? |
| 34 | + answer: 13 billion |
| 35 | + - question: Do all Granite models have the same number of model parameters? |
| 36 | + answer: Models vary from 3 to 34 billion parameters. |
| 37 | + - context: > |
| 38 | + On May 6, 2024, IBM released the source code of four variations of Granite |
| 39 | + Code Models under Apache 2, an open source permissive license that allows |
| 40 | + completely free use, modification and sharing of the software, and put |
| 41 | + them on Hugging Face for public use.[14][15] According to IBM's own |
| 42 | + report, Granite 8b outperforms Llama 3 on several coding related tasks |
| 43 | + within similar range of parameters. |
| 44 | + questions_and_answers: |
| 45 | + - question: When did IBM first release the source code of Granite Code Models? |
| 46 | + answer: May 6, 2024 |
| 47 | + - question: Does Granite outperform other LLMs? |
| 48 | + answer: >- |
| 49 | + Granite 8b outperforms Llama 3 on several coding related tasks within |
| 50 | + similar range of parameters |
| 51 | + - question: Where are the Granite models and code accessed? |
| 52 | + answer: IBM put them on Hugging Face for public use. |
| 53 | + - context: >- |
| 54 | + A foundation model is an AI model trained on broad data at scale such that |
| 55 | + it can be adapted to a wide range of downstream tasks |
| 56 | + questions_and_answers: |
| 57 | + - question: what is a foundation model? |
| 58 | + answer: 'A foundation model is an AI model trained on broad data ' |
| 59 | + - question: What scale of data is a foundation model trained on? |
| 60 | + answer: >- |
| 61 | + the data is at scale such that it can be adapted to a wide range of |
| 62 | + downstream tasks |
| 63 | + - question: why are these tasks referred to as downstream? |
| 64 | + answer: >- |
| 65 | + "Downstream" means "derived from" or "based on"; downstream tasks are |
| 66 | + work that is done using the foundation model. |
| 67 | + - context: >- |
| 68 | + IBM opened the source code of some code models.[8][9] Granite models are |
| 69 | + trained on datasets curated from Internet, academic publishings, code |
| 70 | + datasets, legal and finance documents |
| 71 | + questions_and_answers: |
| 72 | + - question: Did IBM release the full source code for IBM Granite? |
| 73 | + answer: IBM released source code of four code models. |
| 74 | + - question: What datasets is Granite trained on? |
| 75 | + answer: >- |
| 76 | + As well as datasets curated from the internet, Granite is trained on |
| 77 | + academic publications, code datasets, legal and finance documents |
| 78 | + - question: Are the Granite training datasets open source? |
| 79 | + answer: 'Yes' |
| 80 | +document: |
| 81 | + repo: https://github.com/barny/taxonomy-knowledge-docs |
| 82 | + commit: 260ddda8c25526f200a5fe60e3898cf303e2430c |
| 83 | + patterns: |
| 84 | + - granite-20250123T172453388.md |
0 commit comments