Skip to content

Commit 8576128

Browse files
committed
IBM Granite Knowledge
Signed-off-by: I Barnard <[email protected]>
1 parent dfa3afa commit 8576128

File tree

2 files changed

+89
-0
lines changed

2 files changed

+89
-0
lines changed

knowledge/technology/attribution.txt

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,5 @@
1+
Title of work: IBM Granite
2+
Link to work: https://en.wikipedia.org/wiki/IBM_Granite
3+
Revision: https://en.wikipedia.org/w/index.php?title=IBM_Granite&oldid=1269265317
4+
License of the work: CC-BY-SA-4.0
5+
Creator names: Wikipedia Staff

knowledge/technology/qna.yaml

Lines changed: 84 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,84 @@
1+
created_by: barny
2+
version: 3
3+
domain: large-language-model
4+
document_outline: Knowledge contribution about the IBM Granite Model
5+
seed_examples:
6+
- context: >-
7+
IBM Granite is a series of decoder-only AI foundation models created by
8+
IBM.[3] It was announced on September 7, 2023,[4][5] and an initial paper
9+
was published 4 days later.[6] Initially intended for use in the IBM's
10+
cloud-based data and generative AI platform Watsonx along with other
11+
models,[7] IBM opened the source code of some code models.[8][9] Granite
12+
models are trained on datasets curated from Internet, academic
13+
publishings, code datasets, legal and finance documents
14+
questions_and_answers:
15+
- question: What is IBM Granite?
16+
answer: >-
17+
IBM Granite is a series of decoder-only AI foundation models created
18+
by IBM.
19+
- question: When as was IBM Granite announced?
20+
answer: September 7, 2023
21+
- question: What's a series of IBM decoder-only AI foundation models?
22+
answer: IBM Granite
23+
- context: >-
24+
Granite's first foundation models were Granite.13b.instruct and
25+
Granite.13b.chat. The "13b" in their name comes from 13 billion, the
26+
amount of parameters they have as models, lesser than most of the larger
27+
models of the time. Later models vary from 3 to 34 billion parameters.
28+
questions_and_answers:
29+
- question: what were the first foundation models in IBm Granite?
30+
answer: Granite.13b.instruct and Granite.13b.chat
31+
- question: >-
32+
How many model parameters do Granite.13b.instruct and Granite.13b.chat
33+
have?
34+
answer: 13 billion
35+
- question: Do all Granite models have the same number of model parameters?
36+
answer: Models vary from 3 to 34 billion parameters.
37+
- context: >
38+
On May 6, 2024, IBM released the source code of four variations of Granite
39+
Code Models under Apache 2, an open source permissive license that allows
40+
completely free use, modification and sharing of the software, and put
41+
them on Hugging Face for public use.[14][15] According to IBM's own
42+
report, Granite 8b outperforms Llama 3 on several coding related tasks
43+
within similar range of parameters.
44+
questions_and_answers:
45+
- question: When did IBM first release the source code of Granite Code Models?
46+
answer: May 6, 2024
47+
- question: Does Granite outperform other LLMs?
48+
answer: >-
49+
Granite 8b outperforms Llama 3 on several coding related tasks within
50+
similar range of parameters
51+
- question: Where are the Granite models and code accessed?
52+
answer: IBM put them on Hugging Face for public use.
53+
- context: >-
54+
A foundation model is an AI model trained on broad data at scale such that
55+
it can be adapted to a wide range of downstream tasks
56+
questions_and_answers:
57+
- question: what is a foundation model?
58+
answer: 'A foundation model is an AI model trained on broad data '
59+
- question: What scale of data is a foundation model trained on?
60+
answer: >-
61+
the data is at scale such that it can be adapted to a wide range of
62+
downstream tasks
63+
- question: why are these tasks referred to as downstream?
64+
answer: >-
65+
"Downstream" means "derived from" or "based on"; downstream tasks are
66+
work that is done using the foundation model.
67+
- context: >-
68+
IBM opened the source code of some code models.[8][9] Granite models are
69+
trained on datasets curated from Internet, academic publishings, code
70+
datasets, legal and finance documents
71+
questions_and_answers:
72+
- question: Did IBM release the full source code for IBM Granite?
73+
answer: IBM released source code of four code models.
74+
- question: What datasets is Granite trained on?
75+
answer: >-
76+
As well as datasets curated from the internet, Granite is trained on
77+
academic publications, code datasets, legal and finance documents
78+
- question: Are the Granite training datasets open source?
79+
answer: 'Yes'
80+
document:
81+
repo: https://github.com/barny/taxonomy-knowledge-docs
82+
commit: 260ddda8c25526f200a5fe60e3898cf303e2430c
83+
patterns:
84+
- granite-20250123T172453388.md

0 commit comments

Comments
 (0)