Thanks for your wonderful contribution for embedding netflix item data.
In python, when I load your Netflix data, the text_feat.npy and image_feat.npy both represents a numpy adarray. To be more exact:
text_feat = np.load('text_feat.npy')
image_feat = np.load('image_feat.npy')
print(text_feat.shape) # -> 17366 * 768
print(image_feat.shape) # -> 17366 * 512
May I ask if it is true that the organization of text_feat and image_feat are by the sequence of, for each row,
item 1, [embedding 1];
item 2,[embedding 2]; # as itemid sequence
...
or
item 9733, [embedding 9733];
item 14147, [embedding 14147]; # as the sequence from item_attribute.csv
...
Thanks! I am carrying out embedding_based i2i similarity recommendation.
Thanks for your wonderful contribution for embedding netflix item data.
In python, when I load your Netflix data, the
text_feat.npyandimage_feat.npyboth represents a numpy adarray. To be more exact:May I ask if it is true that the organization of text_feat and image_feat are by the sequence of, for each row,
item 1, [embedding 1];
item 2,[embedding 2]; # as itemid sequence
...
or
item 9733, [embedding 9733];
item 14147, [embedding 14147]; # as the sequence from
item_attribute.csv...
Thanks! I am carrying out embedding_based i2i similarity recommendation.