You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+21-17Lines changed: 21 additions & 17 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,10 +4,10 @@
4
4
5
5
> **Note:** All code comments and docstrings are in English for international collaboration and code clarity. All user-facing messages and buttons are automatically translated to the user's selected language.
6
6
7
-
## 🚀 What's New in v4.0.0
7
+
## 🚀 What's New in v4.1.0
8
8
9
9
-**🆕 Multi-Level LLM Architecture**: OpenAI → Groq → Local LLM → Fallback Plan
10
-
-**🆕 Local LLM Integration**: TinyLlama 1.1B model for offline operation
10
+
-**🆕 Local LLM Integration**: Google Gemma 4 model for offline operation
11
11
-**🆕 Guaranteed Availability**: Bot works even without internet connection
12
12
-**🆕 Enhanced Fallback System**: Robust error handling and service switching
13
13
-**🆕 Improved Plan Quality**: Professional-grade study plan templates
@@ -37,7 +37,7 @@ The bot features a sophisticated 4-tier fallback system that ensures reliable se
37
37
|----------|---------|-------------|----------|
38
38
|**1**|**OpenAI GPT**| Primary model for high-quality plans | Best quality, when available |
39
39
|**2**|**Groq**| Secondary model, OpenAI alternative | Fast fallback, reliable service |
40
-
|**3**|**Local LLM**|TinyLlama 1.1B local model | Offline operation, privacy |
40
+
|**3**|**Local LLM**|Google Gemma 4 local model | Offline operation, privacy |
41
41
|**4**|**Fallback Plan**| Predefined professional template | Guaranteed availability |
42
42
43
43
### ⚡ How It Works
@@ -46,7 +46,7 @@ The bot automatically attempts to generate study plans using available services
46
46
47
47
1.**Primary**: OpenAI API (if `OPENAI_API_KEY` is set and quota available)
48
48
2.**Fallback 1**: [Groq](https://groq.com/) (if `GROQ_API_KEY` is set)
49
-
3.**Fallback 2**: Local LLM (TinyLlama 1.1B model)
49
+
3.**Fallback 2**: Local LLM (Google Gemma 4 model)
50
50
4.**Last Resort**: Local plan generator (comprehensive template)
-**Multi-Level LLM Architecture**: OpenAI → Groq → Local LLM → Fallback Plan
243
-
-**Local LLM Integration**: TinyLlama 1.1B model for offline operation
247
+
-**Local LLM Integration**: Google Gemma 4 model for offline operation
244
248
-**Guaranteed Availability**: Bot works even without internet connection
245
249
-**Enhanced Fallback System**: Robust error handling and service switching
246
250
@@ -309,4 +313,4 @@ MIT License - see [LICENSE](LICENSE) file for details.
309
313
310
314
---
311
315
312
-
**EduPlannerBotAI v4.0.0** represents a significant milestone, transforming the bot from a simple OpenAI-dependent service into a robust, enterprise-grade system with guaranteed availability and offline operation capabilities. This release sets the foundation for future enhancements while maintaining backward compatibility and improving overall user experience.
316
+
**EduPlannerBotAI v4.1.0** represents a significant milestone, transforming the bot from a simple OpenAI-dependent service into a robust, enterprise-grade system with guaranteed availability and offline operation capabilities. This release sets the foundation for future enhancements while maintaining backward compatibility and improving overall user experience.
2.**Mistral 7B**: Excellent performance, medium size (~4GB)
108
-
3.**Phi-2**: Good quality, smaller size (~1.4GB)
109
-
110
-
**Note**: Update the model path in `services/local_llm.py` if using a different model.
111
-
112
-
## Performance Tips
113
-
114
-
-**First Run**: The first request will be slower as the model loads into memory
115
-
-**Subsequent Requests**: Much faster after initial loading
116
-
-**Memory**: Keep at least 2GB RAM free for optimal performance
117
-
-**CPU**: Multi-core processors will improve inference speed
118
-
119
-
## Support
120
-
121
-
If you encounter issues with the local LLM:
122
50
123
-
1.Check the bot logs for detailed error messages
124
-
2. Verify the model file is correctly placed
125
-
3. Ensure sufficient system resources
126
-
4. Open an issue on GitHub with error details
51
+
Check:
52
+
-file path in `LOCAL_LLM_MODEL_PATH`
53
+
- read permissions for the model file
54
+
- available RAM/CPU resources
127
55
128
-
##License
56
+
### Out-of-memory or slow responses
129
57
130
-
The TinyLlama model is licensed under Apache 2.0. See the [Hugging Face page](https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF) for full license details.
0 commit comments