skip to main content
LOTERRE

LOTERRE

Search from vocabulary

Content language

| español français
Search help

Concept information

Preferred term

Megatron-LM  

Definition(s)

  • A language model pre-training based on GPU parallelism.

Broader concept(s)

Bibliographic citation(s)

  • • Narayanan, D., Shoeybi, M., Casper, J., LeGresley, P., Patwary, M., Korthikanti, V. A., Vainbrand, D., Kashinkunti, P., Bernauer, J., Catanzaro, B., Phanishayee, A., & Zaharia, M. (2021). Efficient large-scale language model training on GPU clusters using megatron-lm. ArXiv:2104.04473 [Cs]. http://arxiv.org/abs/2104.04473
  • • Shoeybi, M., Patwary, M., Puri, R., LeGresley, P., Casper, J., & Catanzaro, B. (2019). Megatron-LM: Training multi-billion parameter language models using GPU model parallelism. https://arxiv.org/abs/1909.08053v1

base of

has application field

has for input language

is encoded in

is executed in

In other languages

URI

http://data.loterre.fr/ark:/67375/LTK-DCJM3LC1-6

Download this concept:

RDF/XML TURTLE JSON-LD Last modified 6/20/24