FashionM3 / README.md
pangkaicheng
first commit to huggingface
122ccf7
# FashionM3 Project Setup Guide
This guide explains how to set up and run the FashionM3 project, which requires the FashionRec dataset, environment configuration, and specific server startup steps.
**Important Note on Hugging Face Space Deployment:**
Due to the high computational requirements of the underlying Vision-Language Model (VLM) and associated services, this Hugging Face Space currently serves primarily as a **code repository**. The full interactive demo **cannot be run directly on the free tier of Hugging Face Spaces**. Please follow the instructions below to set up and run the FashionM3 application locally on your machine.
## Project Overview
The work is introduced in this paper FashionM3: Multimodal, Multitask, and Multiround Fashion Assistant based on Unified Vision-Language Model - https://arxiv.org/abs/2504.17826
## Prerequisites
- Python 3.10 or higher
- `pip` for installing dependencies
- A working proxy server (if needed, e.g., at `http://127.0.0.1:10809`)
## Step 1: Download the FashionRec Dataset
1. Download the FashionRec dataset from https://huggingface.co/datasets/Anony100/FashionRec
2. Extract the dataset to a directory of your choice (e.g., `/path/to/FashionRec`).
3. Note the absolute path to the dataset directory, as it will be used in the `.env` file.
## Step 2: Configure the Environment File
Create a `.env` file in the project root directory with the following content:
```plaintext
# Chainlit server port
CHAINLIT_PORT=8888
# Proxy configuration (update or remove if not needed)
PROXY=http://127.0.0.1:10809
# API keys (replace with your own keys)
OPENAI_API_KEY=your_openai_api_key_here
GEMINI_API_KEY=your_gemini_api_key_here
# Path to the FashionRec dataset (update to your dataset path)
FASHION_DATA_ROOT=/path/to/FashionRec
# Directory for generated images
GEN_IMG_DIR=./generated_images
```
### Notes:
- Replace `your_openai_api_key_here` and `your_gemini_api_key_here` with your actual OpenAI and Google Gemini API keys.
- Set FASHION_DATA_ROOT to the absolute path of the FashionRec dataset (e.g., /home/user/data/FashionRec).
- Update PROXY to match your proxy server, or remove it if no proxy is used.
## Step 3: Install Dependencies
Install the required Python packages:
```bash
pip install -r requirements.txt
```
## Step 4: Run the Application
Follow these steps to start the FashionM3 application:
### 1. Start the Fashion VLM MCP Server:
Run the MCP server for the fashion vision-language model:
```bash
python mcp_servers/fashion_vlm/main.py
```
Ensure the server starts successfully and remains running.
### 2. Start the FashionM3 Client:
Launch the Chainlit client to interact with the Fashion Assistant:
```bash
chainlit run chainlit_app.py --port 8888
```
### 3. Interact with the Fashion Assistant:
Open your browser and navigate to:
```plaintext
http://localhost:8888/
```
This will load the FashionM3 interface, allowing you to interact with the Fashion Assistant.
## Citation
If you find this work helpful, please consider citing our paper:
```
@article{pang2025fashionm3,
title={FashionM3: Multimodal, Multitask, and Multiround Fashion Assistant based on Unified Vision-Language Model},
author={Pang, Kaicheng and Zou, Xingxing and Wong, Waikeung},
journal={arXiv preprint arXiv:2504.17826},
year={2025}
}
```