2 datasets found

Tags: mT5

Filter Results
  • mT5

    A multilingual version of the seq2seq architecture trained on Colossal Clean Crawled Corpus.
  • mC4

    Parameter-efficient fine-tuning (PEFT) using labeled task data can significantly improve the performance of large language models (LLMs) on the downstream task. However, there...
You can also access this registry using the API (see API Docs).