NJUyued commited on
Commit
ca17441
1 Parent(s): 6905d21

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -3
README.md CHANGED
@@ -1,3 +1,41 @@
1
- ---
2
- license: cc-by-nc-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ ---
4
+
5
+ # PC2-NoiseofWeb
6
+
7
+ This repo releases data introduced in our paper
8
+
9
+ > ***PC2: Pseudo-Classification Based Pseudo-Captioning for Noisy Correspondence Learning in Cross-Modal Retrieval***
10
+ > ***Authors**: Yue Duan, Zhangxuan Gu, Zhenzhe Ying, Lei Qi, Changhua Meng and Yinghuan Shi*
11
+
12
+ Quick links: [[arXiv (coming soon)]() | [Published paper (coming soon)]() | [Poster (coming soon)]() | [Zhihu (coming soon)]() | [Code download]() | [Dataset download](https://drive.google.com/file/d/1MsR9GmRDUj4NoeL4xL8TXpes51JnpsrZ/view?usp=drive_link)]
13
+
14
+ ## Data Collection
15
+ We develop a new dataset named **Noise of Web (NoW)** for NCL. It contains **100K website image-meta description pairs** (**98,000 pairs for training, 1,000 for validation, and 1,000 for testing**), which are open-sourced and can be crawled by anyone. NoW has two main characteristics: *without human annotations and the noisy pairs are naturally captured*. The source data of NoW is obtained by taking screenshots when accessing web pages on mobile devices (resolution: 720 $\times$ 1280) and parsing meta descriptions in html source code. In [NCR](https://github.com/XLearning-SCU/2021-NeurIPS-NCR) (predecessor of NCL), each image in all datasets were preprocessed using Faster-RCNN detector provided by [Bottom-up Attention Model](https://github.com/peteanderson80/bottom-up-attention) to generate 36 region proposals, and each proposal was encoded as a 2048-dimensional feature. Thus, following NCR, we release our the features instead of raw images for fair comparison. However, we can not just use detection methods like Faster-RCNN to extract image features since it is trained on real-world animals and objects on MS-COCO. To tackle this, we adapt [APT](https://openaccess.thecvf.com/content/CVPR2023/papers/Gu_Mobile_User_Interface_Element_Detection_via_Adaptively_Prompt_Tuning_CVPR_2023_paper.pdf) as the detection model since it is trained on the mobile user interface data. Similar to existing datasets, we capture top 36 objects with their features for one image, that is, we can encode one image into a 36 $\times$ 768 matrix. We do not artificially control the noise ratio, as all data is obtained automatically and randomly over the web. **The estimated noise ratio of this dataset is nearly 70%**. Due to the automated and non-human curated data collection process, the noise in NoW is highly authentic and intrinsic.
16
+
17
+ ## Data Structure
18
+
19
+ ```
20
+
21
+ |-- h5100k_precomp
22
+ | |-- dev_caps_bpe.txt
23
+ | |-- dev_caps_bert.txt
24
+ | |-- dev_ids.txt
25
+ | |-- dev_ims.npy
26
+ | |-- test_caps_bpe.txt
27
+ | |-- test_caps_bert.txt
28
+ | |-- test_ids.txt
29
+ | |-- test_ims.npy
30
+ | |-- train_caps_bpe.txt
31
+ | |-- train_caps_bert.txt
32
+ | |-- train_ids.txt
33
+ | |-- train_ims.npy
34
+ |-- vocab
35
+ | |-- now100k_precomp_vocab_bert.json
36
+ | |-- now100k_precomp_vocab_bpe.json
37
+ | |-- now100k_precomp_vocab_jieba.json
38
+
39
+ ```
40
+
41
+ Please note that since our raw data contains some sensitive business data, we only provide the **encoded image features** (\*_ims.npy) and the **token ids of the text tokenized**. For tokenizer, we use both [Tokenizers](https://github.com/huggingface/tokenizers) with [BPE](https://huggingface.co/docs/tokenizers/api/models#tokenizers.models.BPE) to produce \*_caps_bpe.txt and [BertTokenizer](https://huggingface.co/transformers/v3.0.2/model_doc/bert.html#berttokenizer) with [bert-base-multilingual-cased](https://huggingface.co/google-bert/bert-base-multilingual-cased) pre-trained model to produce \*_caps_bert.txt. **Our vocabulary size of BPE tokenizer is 10,000 and that of BertTokenizer is 32702**. \*_ids.txt records the serial number of the data in the original 500k dataset. In the future, we may process and make the original dataset public.