personainsight / app.py
dindizz's picture
Update app.py
2875e4e verified
import gradio as gr
from transformers import CLIPProcessor, CLIPModel
from PIL import Image
import torch
# Load a pre-trained CLIP model and processor from Hugging Face
model_name = "openai/clip-vit-base-patch32"
model = CLIPModel.from_pretrained(model_name)
processor = CLIPProcessor.from_pretrained(model_name)
# Define possible personas mapped to interests
predefined_personas = {
"Fashion": {
"Age": "18-30 years",
"Gender": "Likely Female",
"Interests": ["Fashion", "Trendy Clothing", "Social Media"],
"Income Level": "Medium",
"Psychographics": ["Fashion-conscious", "Socially active", "Trend-seeker"],
"Behavioral Traits": ["Follows fashion influencers", "Buys seasonal items", "Active on Instagram"]
},
"Technology": {
"Age": "20-40 years",
"Gender": "Likely Male",
"Interests": ["Technology", "Gadgets", "Software"],
"Income Level": "High",
"Psychographics": ["Tech-savvy", "Innovative", "Early adopter"],
"Behavioral Traits": ["Follows tech blogs", "Interested in product launches", "Buys online frequently"]
},
"Sports": {
"Age": "15-40 years",
"Gender": "Likely Male",
"Interests": ["Sports", "Fitness", "Outdoor Activities"],
"Income Level": "Medium",
"Psychographics": ["Active lifestyle", "Health-conscious", "Competitive"],
"Behavioral Traits": ["Watches sports events", "Goes to the gym", "Buys athletic gear"]
},
"Travel": {
"Age": "25-45 years",
"Gender": "Likely Female or Male",
"Interests": ["Travel", "Adventure", "Culture"],
"Income Level": "Medium to High",
"Psychographics": ["Experience-seeker", "Open-minded", "Curious"],
"Behavioral Traits": ["Follows travel blogs", "Likes outdoor activities", "Prefers unique experiences"]
},
"Luxury": {
"Age": "30-50 years",
"Gender": "Likely Female or Male",
"Interests": ["Luxury Goods", "High-end Fashion", "Fine Dining"],
"Income Level": "High",
"Psychographics": ["Status-conscious", "Exclusive-taste", "Prestige-oriented"],
"Behavioral Traits": ["Buys luxury brands", "Frequent traveler", "Invests in high-end products"]
},
"Fitness": {
"Age": "20-40 years",
"Gender": "Likely Female or Male",
"Interests": ["Fitness", "Healthy Living", "Nutrition"],
"Income Level": "Medium",
"Psychographics": ["Health-conscious", "Goal-oriented", "Disciplined"],
"Behavioral Traits": ["Goes to the gym regularly", "Buys fitness equipment", "Follows nutrition plans"]
},
"Education": {
"Age": "18-35 years",
"Gender": "Likely Female or Male",
"Interests": ["Education", "Learning", "Career Development"],
"Income Level": "Medium",
"Psychographics": ["Growth-oriented", "Knowledge-seeker", "Curious"],
"Behavioral Traits": ["Follows online courses", "Interested in higher education", "Learns new skills"]
},
"Entertainment": {
"Age": "15-35 years",
"Gender": "Likely Female or Male",
"Interests": ["Movies", "Music", "Video Games"],
"Income Level": "Medium",
"Psychographics": ["Fun-seeking", "Relaxed", "Entertainment-driven"],
"Behavioral Traits": ["Watches Netflix", "Plays video games", "Follows pop culture trends"]
}
}
# List of interest categories to compare against
interest_categories = list(predefined_personas.keys())
# Function to calculate similarity between the image and each interest category
def get_most_similar_persona(image):
"""Infers the most similar predefined persona for the given image."""
inputs = processor(images=image, return_tensors="pt")
with torch.no_grad():
image_features = model.get_image_features(**inputs)
# Define descriptions of interest categories (texts)
category_descriptions = [
f"A photo that represents {category.lower()}." for category in interest_categories
]
# Tokenize the category descriptions and get text features
text_inputs = processor(text=category_descriptions, return_tensors="pt", padding=True)
text_features = model.get_text_features(**text_inputs)
# Calculate similarity between image and text features
image_features /= image_features.norm(dim=-1, keepdim=True)
text_features /= text_features.norm(dim=-1, keepdim=True)
similarities = torch.matmul(image_features, text_features.T)
# Find the most similar category
most_similar_idx = similarities.argmax().item()
most_similar_category = interest_categories[most_similar_idx]
return predefined_personas[most_similar_category]
# Gradio interface function
def persona_analysis(image):
"""Generates a marketing persona based on the image."""
persona = get_most_similar_persona(image)
return "\n".join([f"{key}: {value}" for key, value in persona.items()])
# Build the Gradio interface
iface = gr.Interface(
fn=persona_analysis,
inputs=gr.Image(type="pil"), # Accept image input in PIL format
outputs=gr.Textbox(), # Display persona as text output
title="Marketing Persona Generator",
description="Upload an image to generate a marketing persona based on the ad subject."
)
# Launch the Gradio interface
iface.launch()