familidata_gemma7b / README.md
hyokwan's picture
Update README.md
e8dbabc verified
metadata
license: mit
datasets:
  - hyokwan/common
language:
  - ko
metrics:
  - accuracy
base_model:
  - google/gemma-7b-it
pipeline_tag: text-generation
library_name: transformers

Model Details model

model is continued pretrained language model based on google/gemma-7b-it

This model is trained with specific department of university (Korea Plytechnics Fintech) .

Intended Use TBD

How to use TBD

Responsibility & Safety We believe that an open approach to AI leads to better, safer products, faster innovation, and a bigger overall market. We are committed to Responsible AI development and took a series of steps to limit misuse and harm and support the open source community.

Foundation models are widely capable technologies that are built to be used for a diverse range of applications. They are not designed to meet every developer preference on safety levels for all use cases, out-of-the-box, as those by their nature will differ across different applications.

Rather, responsible LLM-application deployment is achieved by implementing a series of safety best practices throughout the development of such applications, from the model pre-training, fine-tuning and the deployment of systems composed of safeguards to tailor the safety needs specifically to the use case and audience.

As part of the Llama 3 release, we updated our Responsible Use Guide to outline the steps and best practices for developers to implement model and system level safety for their application. We also provide a set of resources including Meta Llama Guard 2 and Code Shield safeguards. These tools have proven to drastically reduce residual risks of LLM Systems, while maintaining a high level of helpfulness. We encourage developers to tune and deploy these safeguards according to their needs and we provide a reference implementation to get you started.

Responsible release In addition to responsible use considerations outlined above, we followed a rigorous process that requires us to take extra measures against misuse and critical risks before we make our release decision.

Misuse


license: mit datasets: - hyokwan/common language: - ko metrics: - accuracy base_model: - google/gemma-7b-it pipeline_tag: text-generation library_name: transformers tags: - finance - education