OPT

Description of OPT

OPT (Open Pre-trained Transformer) is a library of open language models from Meta AI, ranging from 125M to 175B parameters, developed as a transparent alternative to GPT-3. The models were trained on approximately 180 billion English tokens, and along with the weights, the code and a training logbook were published for reproducibility and analysis. OPT-175B is comparable in quality to GPT-3 while being created with a smaller carbon footprint and a focus on accessibility for the research community. The OPT family includes lightweight variants for embedding into products and large models for complex reasoning, text generation, and code generation tasks. OPT can be used to build research platforms, assistant prototypes, experiments with RLHF and instruction fine-tuning, as well as enterprise chatbots and analytics services within the customer’s infrastructure. FreeBlock helps select the right OPT size, fine-tune the model for your tasks, build RAG and agent pipelines, and deploy them securely on your servers or in the cloud. If you want to combine the openness and power of Meta models, order AI project development on OPT from FreeBlock.

Other technologies

Submit an application

!
The field is filled in incorrectly
!
The field is filled in incorrectly
Мы обрабатываются файлы cookie. Оставаясь на сайте, вы даёте своё согласие на использование cookie в соответствии с политикой конфиденциальности