Skip to content

Commit 4383e36

Browse files
committed
adding more info about the distilled model
1 parent a171b25 commit 4383e36

1 file changed

Lines changed: 11 additions & 2 deletions

File tree

rag_and_distilled_model/README.md

Lines changed: 11 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -41,9 +41,18 @@ JSON file.
4141

4242
## Models Used
4343

44-
* **LaMini-Flan-T5-248M**: It is a Local LLM and it is a distilled version of
45-
Google's Flan-T5, optimized for lightweight text generation tasks.
44+
* **LaMini-Flan-T5-248M**: This model is one of the LaMini-LM model series in
45+
paper "LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale
46+
Instructions". It is a Local LLM and it is a distilled version of
47+
Google's Flan-T5, fine-tuned on LaMini-instruction dataset for instruction
48+
fine-tuning, it's optimized for lightweight text generation tasks.
4649
Used here for reasoning, summarization, and RAG response generation.
50+
Its knowledge is distilled from large language models by performing
51+
sentence/offline distillation (Kim and Rush, 2016). A total of
52+
2.58M pairs of instructions and responses was generated using gpt-3.5-turbo
53+
based on several existing resources of prompts, including self-instruct
54+
(Wang et al., 2022), P3 (Sanh et al., 2022), Flan (Longpre et al., 2023) and
55+
Alpaca (Taori et al., 2023).
4756
* **all-MiniLM-L6-v2**: It as an Embedding model and it is a compact sentence-transformer
4857
model used to convert text chunks into numerical vector embeddings for
4958
semantic search and retrieval.

0 commit comments

Comments
 (0)