Update README.md
Browse files
README.md
CHANGED
@@ -8,6 +8,7 @@ tags:
|
|
8 |
pipeline_tag: text-to-video
|
9 |
---
|
10 |
|
|
|
11 |
# Model Card
|
12 |
## Details
|
13 |
This model underwent training using CLIP4Clip, a video retrieval method based on the CLIP framework, as described in the paper [CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip
|
@@ -17,6 +18,7 @@ The training process involved 150,000 videos obtained from the [WebVid Dataset](
|
|
17 |
|
18 |
To adapt the clip model obtained during training into the implementation of [clip-vit-base-patch32](https://huggingface.co/openai/clip-vit-base-patch32), we have made modifications to the weights.
|
19 |
|
|
|
20 |
### Use with Transformers
|
21 |
### Extracting Text Embeddings:
|
22 |
|
@@ -41,7 +43,9 @@ print("sequence_output: ", sequence_output)
|
|
41 |
```
|
42 |
|
43 |
### Extracting Video Embeddings:
|
44 |
-
|
|
|
|
|
45 |
|
46 |
## Model Intended Use
|
47 |
|
|
|
8 |
pipeline_tag: text-to-video
|
9 |
---
|
10 |
|
11 |
+
|
12 |
# Model Card
|
13 |
## Details
|
14 |
This model underwent training using CLIP4Clip, a video retrieval method based on the CLIP framework, as described in the paper [CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip
|
|
|
18 |
|
19 |
To adapt the clip model obtained during training into the implementation of [clip-vit-base-patch32](https://huggingface.co/openai/clip-vit-base-patch32), we have made modifications to the weights.
|
20 |
|
21 |
+
|
22 |
### Use with Transformers
|
23 |
### Extracting Text Embeddings:
|
24 |
|
|
|
43 |
```
|
44 |
|
45 |
### Extracting Video Embeddings:
|
46 |
+
|
47 |
+
An additional [notebook](https://huggingface.co/Diangle/clip4clip-webvid/blob/main/Notebooks/GSI_VideoRetrieval_EmbedVideos.ipynb) is available that provides instructions on how to perform video embedding.
|
48 |
+
|
49 |
|
50 |
## Model Intended Use
|
51 |
|