Skip to content

Commit

Permalink
Update doc: change LLaMA as Prototype
Browse files Browse the repository at this point in the history
  • Loading branch information
aws-bowencc authored and hannanjgaws committed Jul 21, 2023
1 parent 18b448f commit ea9dd5f
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -419,7 +419,7 @@ for running HuggingFace `facebook/opt-13b` autoregressive sampling on a trn1.2xl
- [OPT](https://huggingface.co/docs/transformers/model_doc/opt)
- [GPT-Neox [Experimental]](https://huggingface.co/docs/transformers/model_doc/gpt_neox)
- [Bloom [Experimental]](https://huggingface.co/docs/transformers/model_doc/bloom)
- [LLaMA [Experimental]](https://huggingface.co/docs/transformers/main/model_doc/llama)
- [LLaMA [Prototype]](https://huggingface.co/docs/transformers/main/model_doc/llama)

# Upcoming features

Expand Down
2 changes: 1 addition & 1 deletion releasenotes.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ Date: 2023-07-03

- [Experimental] Added support for GPT-NeoX models.
- [Experimental] Added support for BLOOM models.
- [Experimental] Added support for LLaMA models.
- [Prototype] Added support for LLaMA models.
- Added support for more flexible tensor-parallel configurations to GPT2, OPT, and BLOOM. Previously, we had two constraints on `tp_degree`: 1) The attention heads needs to be evenly divisible by `tp_degree` 2) The `tp_degree` needs to satisfy the runtime topologies constraint for collective communication (i.e Allreduce). For more details on supported topologies, see: [Tensor-parallelism support](README.md#tensor-parallelism-support) and https://awsdocs-neuron.readthedocs-hosted.com/en/latest/general/arch/neuron-features/collective-communication.html. We now get rid of 1) by using 1-axis padding.
- Added multi-query / multi-group attention support for GPT2.

Expand Down

0 comments on commit ea9dd5f

Please sign in to comment.