Large Language Model (LLM) Training Algorithm: Difference between revisions

From GM-RKB
Jump to navigation Jump to search
 
Line 87: Line 87:


=== 2025 ===
=== 2025 ===
== References ==
* ([[Kumar et al., 2025]]) ⇒ [[Komal Kumar]], [[Tajamul Ashraf]], [[Omkar Thawakar]], [[Rao Muhammad Anwer]], [[Hisham Cholakkal]], [[Mubarak Shah]], [[Ming-Hsuan Yang]], [[Phillip H. S. Torr]], [[Salman Khan]], and [[Fahad Shahbaz Khan]]. ([[2025]]). “LLM Post-Training: A Deep Dive Into Reasoning Large Language Models.”  [http://dx.doi.org/10.48550/arXiv.2502.21321 doi:10.48550/arXiv.2502.21321]  
* ([[Kumar et al., 2025]]) ⇒ [[Komal Kumar]], [[Tajamul Ashraf]], [[Omkar Thawakar]], [[Rao Muhammad Anwer]], [[Hisham Cholakkal]], [[Mubarak Shah]], [[Ming-Hsuan Yang]], [[Phillip H. S. Torr]], [[Salman Khan]], and [[Fahad Shahbaz Khan]]. ([[2025]]). “LLM Post-Training: A Deep Dive Into Reasoning Large Language Models.”  [http://dx.doi.org/10.48550/arXiv.2502.21321 doi:10.48550/arXiv.2502.21321]  
** NOTES:  
** NOTES:  

Latest revision as of 23:16, 3 March 2025

A Large Language Model (LLM) Training Algorithm is a deep neural model training algorithm that can be implemented by an LLM training system (optimizes large language model parameters) to support LLM training tasks.



References

2025