| |
|
| | import gradio as gr |
| | import clip,torch |
| | import requests |
| | from PIL import Image |
| | import numpy as np |
| | import torch |
| | import torch.nn as nn |
| | from io import BytesIO |
| | import urllib.request |
| |
|
| | |
| | |
| |
|
| | from selenium import webdriver |
| | from selenium.webdriver.common.by import By |
| |
|
| | |
| | def test2(): |
| | driver = webdriver.Chrome() |
| | |
| | driver.get('https://www.hiphoper.com/') |
| | |
| | imgs = driver.find_elements(By.CSS_SELECTOR,'img.card__image') |
| | result = [] |
| | |
| | for img in imgs: |
| | |
| | result.append(img.get_attribute('src')) |
| | |
| | driver.quit() |
| | |
| | return result |
| |
|
| | |
| | def similarity(v1,v2,type=0): |
| | if type ==0: |
| | v1_norm = np.linalg.norm(v1) |
| | v2_norm = np.linalg.norm(v2) |
| | |
| | return np.dot(v1,v2)/(v1_norm*v2_norm) |
| | else: |
| | return np.sqrt(np.sum((v1-v2)**2)) |
| |
|
| | |
| | def democlip(url ,texts): |
| | |
| | if url =='': |
| | print('SYSTEM : alternative url') |
| | url = 'https://i.pinimg.com/564x/47/b5/5d/47b55de6f168db65cf46d7d1f0451b64.jpg' |
| | else: |
| | print('SYSTEM : URL progressed') |
| | |
| | if texts =='': |
| | texts ='black desk room girl flower' |
| | else: |
| | print('SYSTEM : TEXT progressed') |
| | |
| | response = requests.get(url) |
| | image_bytes = response.content |
| | texts = list(texts.split(' ')) |
| |
|
| | """Gets the embedding values for the image.""" |
| | device = "cuda" if torch.cuda.is_available() else "cpu" |
| | model, preprocess = clip.load("ViT-B/32", device=device) |
| |
|
| | |
| | text_token = clip.tokenize(texts).to(device) |
| | image = preprocess(Image.open(BytesIO(image_bytes))).unsqueeze(0).to(device) |
| |
|
| | with torch.no_grad(): |
| | image_features = model.encode_image(image) |
| | text_features = model.encode_text(text_token) |
| | |
| | logits_per_image, logits_per_text = model(image,text_token) |
| | probs = logits_per_image.softmax(dim=-1).cpu().numpy() |
| |
|
| | word_dict = {'image':{},'text':{}} |
| | |
| | |
| | for i,text in enumerate(texts): |
| | word_dict['text'][text] = text_features[i].cpu().numpy() |
| | |
| | |
| | for i,img in enumerate(image): |
| | word_dict['image'][img] = image_features[i].cpu().numpy() |
| | |
| | |
| | |
| | tu,ts,tv = torch.pca_lowrank(text_features,center=True) |
| |
|
| | text_pca = torch.matmul(text_features,tv[:,:3]) |
| |
|
| | |
| | imgu,imgs,imgv = torch.pca_lowrank(image_features,center=True) |
| |
|
| | image_pca = torch.matmul(image_features,imgv[:,:3]) |
| |
|
| | |
| | print(text_pca.shape,image_pca.shape) |
| | return text_pca,image_pca |
| |
|
| |
|
| |
|
| | def PCA(img_emb, text_emb,n_components = 3): |
| | x = torch.tensor([[1.,2.,3.,7.],[4.,5.,3.,6.],[7.,9.,8.,9.],[11.,13.,17.,11.]]) |
| | |
| |
|
| | print(x.shape) |
| | u,s,v = torch.pca_lowrank(x,q=None, center=False,niter=2) |
| |
|
| | u.shape,s.shape,v.shape |
| |
|
| | u@torch.diag(s)@v.T |
| |
|
| | |
| | pass |
| |
|
| |
|
| |
|
| | |
| |
|
| | |
| |
|
| | |
| | demo = gr.Interface( |
| | fn=democlip, |
| | |
| | inputs = ['text',gr.Textbox(label='input prediction')], |
| | |
| | outputs = [gr.Textbox(label='text pca Box'),gr.Textbox(label='image pca Box')] |
| | ) |
| | demo.launch() |