You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: model_data.json
+30-30Lines changed: 30 additions & 30 deletions
Original file line number
Diff line number
Diff line change
@@ -965,6 +965,36 @@
965
965
"1.0.0": "Initial release"
966
966
}
967
967
},
968
+
"hf_llama3_vila_m3_3b": {
969
+
"model_name": "VILA_M3_3B",
970
+
"description": "VILA_M3 is a medical vision language model that enhances VLMs with medical expert knowledge, utilizing domain-expert models to improve precision in medical imaging tasks.",
971
+
"authors": "Vishwesh Nath, Wenqi Li, Dong Yang, Andriy Myronenko, et al. from NVIDIA, SingHealth, and NIH",
972
+
"papers": [
973
+
"Nath, Vishwesh, et al. 'VILA_M3: Enhancing Vision-Language Models with Medical Expert Knowledge', arXiv preprint arXiv:2411.12915 (2025)."
974
+
],
975
+
"version": "1.0.0",
976
+
"model_id": "hf_llama3_vila_m3_3b",
977
+
"readme": "<h1>VILA_M3_3B</h1>\n<p>VILA_M3 is a medical vision language model that enhances VLMs with medical expert knowledge, utilizing domain-expert models to improve precision in medical imaging tasks.</p>\n<p>This model is available at: <a href=\"https://huggingface.co/MONAI/Llama3-VILA-M3-3B\">MONAI/Llama3-VILA-M3-3B</a></p>\n<h2>Citation</h2>\n<pre><code>@article{nath2025vila,\n title={VILA_M3: Enhancing Vision-Language Models with Medical Expert Knowledge},\n author={Nath, Vishwesh and Li, Wenqi and Yang, Dong and Myronenko, Andriy and Zheng, Mingxin and Lu, Yao and Liu, Zhijian and Yin, Hongxu and Tang, Yucheng and Guo, Pengfei and Zhao, Can and Xu, Ziyue and He, Yufan and Law, Yee Man and Simon, Benjamin and Harmon, Stephanie and Heinrich, Greg and Aylward, Stephen and Edgar, Marc and Zephyr, Michael and Han, Song and Molchanov, Pavlo and Turkbey, Baris and Roth, Holger and Xu, Daguang},\n journal={arXiv preprint arXiv:2411.12915},\n year={2025}\n}\n</code></pre>",
"description": "VILA_M3 is a medical vision language model that enhances VLMs with medical expert knowledge, utilizing domain-expert models to improve precision in medical imaging tasks.",
986
+
"authors": "Vishwesh Nath, Wenqi Li, Dong Yang, Andriy Myronenko, et al. from NVIDIA, SingHealth, and NIH",
987
+
"papers": [
988
+
"Nath, Vishwesh, et al. 'VILA_M3: Enhancing Vision-Language Models with Medical Expert Knowledge', arXiv preprint arXiv:2411.12915 (2025)."
989
+
],
990
+
"version": "1.0.0",
991
+
"model_id": "hf_llama3_vila_m3_13b",
992
+
"readme": "<h1>VILA_M3_13B</h1>\n<p>VILA_M3 is a medical vision language model that enhances VLMs with medical expert knowledge, utilizing domain-expert models to improve precision in medical imaging tasks.</p>\n<p>This model is available at: <a href=\"https://huggingface.co/MONAI/Llama3-VILA-M3-13B\">MONAI/Llama3-VILA-M3-13B</a></p>\n<h2>Citation</h2>\n<pre><code>@article{nath2025vila,\n title={VILA_M3: Enhancing Vision-Language Models with Medical Expert Knowledge},\n author={Nath, Vishwesh and Li, Wenqi and Yang, Dong and Myronenko, Andriy and Zheng, Mingxin and Lu, Yao and Liu, Zhijian and Yin, Hongxu and Tang, Yucheng and Guo, Pengfei and Zhao, Can and Xu, Ziyue and He, Yufan and Law, Yee Man and Simon, Benjamin and Harmon, Stephanie and Heinrich, Greg and Aylward, Stephen and Edgar, Marc and Zephyr, Michael and Han, Song and Molchanov, Pavlo and Turkbey, Baris and Roth, Holger and Xu, Daguang},\n journal={arXiv preprint arXiv:2411.12915},\n year={2025}\n}\n</code></pre>",
"description": "EXAONEPath is a patch-level pathology pretrained model with 86 million parameters, pretrained on 285,153,903 patches extracted from 34,795 WSIs.",
@@ -1011,35 +1041,5 @@
1011
1041
"changelog": {
1012
1042
"1.0.0": "initial release of CT_CHAT model"
1013
1043
}
1014
-
},
1015
-
"hf_llama3_vila_m3_3b": {
1016
-
"model_name": "VILA_M3_3B",
1017
-
"description": "VILA_M3 is a medical vision language model that enhances VLMs with medical expert knowledge, utilizing domain-expert models to improve precision in medical imaging tasks.",
1018
-
"authors": "Vishwesh Nath, Wenqi Li, Dong Yang, Andriy Myronenko, et al. from NVIDIA, SingHealth, and NIH",
1019
-
"papers": [
1020
-
"Nath, Vishwesh, et al. 'VILA_M3: Enhancing Vision-Language Models with Medical Expert Knowledge', arXiv preprint arXiv:2411.12915 (2025)."
1021
-
],
1022
-
"version": "1.0.0",
1023
-
"model_id": "hf_llama3_vila_m3_3b",
1024
-
"readme": "<h1>VILA_M3_3B</h1>\n<p>VILA_M3 is a medical vision language model that enhances VLMs with medical expert knowledge, utilizing domain-expert models to improve precision in medical imaging tasks.</p>\n<p>This model is available at: <a href=\"https://huggingface.co/MONAI/Llama3-VILA-M3-3B\">MONAI/Llama3-VILA-M3-3B</a></p>\n<h2>Citation</h2>\n<pre><code>@article{nath2025vila,\n title={VILA_M3: Enhancing Vision-Language Models with Medical Expert Knowledge},\n author={Nath, Vishwesh and Li, Wenqi and Yang, Dong and Myronenko, Andriy and Zheng, Mingxin and Lu, Yao and Liu, Zhijian and Yin, Hongxu and Tang, Yucheng and Guo, Pengfei and Zhao, Can and Xu, Ziyue and He, Yufan and Law, Yee Man and Simon, Benjamin and Harmon, Stephanie and Heinrich, Greg and Aylward, Stephen and Edgar, Marc and Zephyr, Michael and Han, Song and Molchanov, Pavlo and Turkbey, Baris and Roth, Holger and Xu, Daguang},\n journal={arXiv preprint arXiv:2411.12915},\n year={2025}\n}\n</code></pre>",
"description": "VILA_M3 is a medical vision language model that enhances VLMs with medical expert knowledge, utilizing domain-expert models to improve precision in medical imaging tasks.",
1033
-
"authors": "Vishwesh Nath, Wenqi Li, Dong Yang, Andriy Myronenko, et al. from NVIDIA, SingHealth, and NIH",
1034
-
"papers": [
1035
-
"Nath, Vishwesh, et al. 'VILA_M3: Enhancing Vision-Language Models with Medical Expert Knowledge', arXiv preprint arXiv:2411.12915 (2025)."
1036
-
],
1037
-
"version": "1.0.0",
1038
-
"model_id": "hf_llama3_vila_m3_13b",
1039
-
"readme": "<h1>VILA_M3_13B</h1>\n<p>VILA_M3 is a medical vision language model that enhances VLMs with medical expert knowledge, utilizing domain-expert models to improve precision in medical imaging tasks.</p>\n<p>This model is available at: <a href=\"https://huggingface.co/MONAI/Llama3-VILA-M3-13B\">MONAI/Llama3-VILA-M3-13B</a></p>\n<h2>Citation</h2>\n<pre><code>@article{nath2025vila,\n title={VILA_M3: Enhancing Vision-Language Models with Medical Expert Knowledge},\n author={Nath, Vishwesh and Li, Wenqi and Yang, Dong and Myronenko, Andriy and Zheng, Mingxin and Lu, Yao and Liu, Zhijian and Yin, Hongxu and Tang, Yucheng and Guo, Pengfei and Zhao, Can and Xu, Ziyue and He, Yufan and Law, Yee Man and Simon, Benjamin and Harmon, Stephanie and Heinrich, Greg and Aylward, Stephen and Edgar, Marc and Zephyr, Michael and Han, Song and Molchanov, Pavlo and Turkbey, Baris and Roth, Holger and Xu, Daguang},\n journal={arXiv preprint arXiv:2411.12915},\n year={2025}\n}\n</code></pre>",
<pclass="text-gray-700">Watch all our working group meetings and discussions on our YouTube playlist. Subscribe to stay updated with the latest developments in federated learning.</p>
0 commit comments