BiliSakura commited on
Commit
78040c4
·
verified ·
1 Parent(s): a639eb3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -48
README.md CHANGED
@@ -1,48 +1,46 @@
1
- ---
2
- license: mit
3
- tags:
4
- - clip
5
- - feature-extraction
6
- ---
7
-
8
- # DGTRS-CLIP-ViT-L-14
9
-
10
- This is the DGTRS-CLIP-ViT-L-14 model. It can be used for a variety of tasks, including zero-shot image classification and text-image retrieval.
11
-
12
- This model is compatible with both the `transformers` and `diffusers` libraries.
13
-
14
- ## How to use
15
-
16
- ### With `transformers`
17
-
18
- ```python
19
- from transformers import CLIPProcessor, CLIPModel
20
-
21
- model = CLIPModel.from_pretrained("BiliSakura/DGTRS-CLIP-ViT-L-14")
22
- processor = CLIPProcessor.from_pretrained("BiliSakura/DGTRS-CLIP-ViT-L-14")
23
-
24
- # Your code here to use the model for image-text similarity, zero-shot classification, etc.
25
- ```
26
-
27
- ### With `diffusers`
28
-
29
- This model's text encoder can be used with Stable Diffusion:
30
-
31
- ```python
32
- # Your code here to use the text encoder with a diffusion model.
33
- ```
34
-
35
- ## Citation
36
-
37
- If you use this model in your research, please cite the original paper:
38
-
39
- ```
40
- @misc{chen2024dual,
41
- title={Dual-granularity Text-Guidance for Text-to-Image Generation},
42
- author={Mitsui Chen and Yiyang Ma and Zesu Liu and Hong-Yu Zhou and Yu-cheng Chen and Jian-wei Liu and Shu-ui Liu and Yu-gang Jiang and Wei-shi Zheng},
43
- year={2024},
44
- eprint={2406.16510},
45
- archivePrefix={arXiv},
46
- primaryClass={cs.CV}
47
- }
48
- ```
 
1
+ ---
2
+ license: mit
3
+ tags:
4
+ - clip
5
+ - feature-extraction
6
+ ---
7
+
8
+ # DGTRS-CLIP-ViT-L-14
9
+
10
+ This is the DGTRS-CLIP-ViT-L-14 model. It can be used for a variety of tasks, including zero-shot image classification and text-image retrieval.
11
+
12
+ This model is compatible with both the `transformers` and `diffusers` libraries.
13
+
14
+ ## How to use
15
+
16
+ ### With `transformers`
17
+
18
+ ```python
19
+ from transformers import CLIPProcessor, CLIPModel
20
+
21
+ model = CLIPModel.from_pretrained("BiliSakura/DGTRS-CLIP-ViT-L-14")
22
+ processor = CLIPProcessor.from_pretrained("BiliSakura/DGTRS-CLIP-ViT-L-14")
23
+
24
+ # Your code here to use the model for image-text similarity, zero-shot classification, etc.
25
+ ```
26
+
27
+ ### With `diffusers`
28
+
29
+ This model's text encoder can be used with Stable Diffusion:
30
+
31
+ ```python
32
+ # Your code here to use the text encoder with a diffusion model.
33
+ ```
34
+
35
+ ## Citation
36
+
37
+ If you use this model in your research, please cite the original paper:
38
+
39
+ ```
40
+ @article{chen2025lrsclip,
41
+ title={LRSCLIP: A Vision-Language Foundation Model for Aligning Remote Sensing Image with Longer Text},
42
+ author={Chen, Weizhi and Chen, Jingbo and Deng, Yupeng and Chen, Jiansheng and Feng, Yuman and Xi, Zhihao and Liu, Diyou and Li, Kai and Meng, Yu},
43
+ journal={arXiv preprint arXiv:2503.19311},
44
+ year={2025}
45
+ }
46
+ ```