--- license: other license_name: chameleon-research-license license_link: https://ai.meta.com/resources/models-and-libraries/chameleon-license/ extra_gated_prompt: '### META CHAMELEON RESEARCH LICENSE AGREEMENT' extra_gated_fields: First Name: text Last Name: text Date of birth: date_picker Country: country Affiliation: text I accept the terms and conditions: checkbox geo: ip_location extra_gated_description: Meta Chameleon Research License and Acceptable Use Policy extra_gated_button_content: I Accept Meta Chameleon Research License and AUP pipeline_tag: image-text-to-text --- # **Meta Chameleon 7B** Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR. See the [Chameleon paper](//arxiv.org/abs/2405.09818) for more information. The [Chameleon collection](//huggingface.co/collections/facebook/chameleon-668da9663f80d483b4c61f58) on HuggingFace contains [7 billion parameter](//huggingface.co/facebook/chameleon-7b) and [30 billion parameter](//huggingface.co/facebook/chameleon-30b) model checkpoints. *[more details and usage examples coming soon]* ### Citation To cite the paper, model, or software, please use the below: ``` @article{Chameleon_Team_Chameleon_Mixed-Modal_Early-Fusion_2024, author = {Chameleon Team}, doi = {10.48550/arXiv.2405.09818}, journal = {arXiv preprint arXiv:2405.09818}, title = {Chameleon: Mixed-Modal Early-Fusion Foundation Models}, url = {https://github.com/facebookresearch/chameleon}, year = {2024} } ``` ### License Use of this repository and related resources are governed by the [Chameleon Research License](//ai.meta.com/resources/models-and-libraries/chameleon-license) and this repository's [LICENSE](./LICENSE) file.