gpt-neo-2.7B-8bit / README.md
skylersterling's picture
Update README.md
68ca011
|
raw
history blame
1.19 kB
metadata
language: en
license: mit
tags:
  - causal-lm
datasets:
  - The_Pile

Quantized EleutherAI/gpt-neo-2.7B with 8-bit weights

This is a version of EleutherAI's GPT-Neo with 2.7 billion parameters that is modified so you can generate and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti). Inspired by GPT-J 8bit.

Here's how to run it: colab

Model Description

GPT-Neo 2.7B is a transformer model designed using EleutherAI's replication of the GPT-3 architecture. GPT-Neo refers to the class of models, while 2.7B represents the number of parameters of this particular pre-trained model.

Links