Linoy Tsaban PRO

linoyts

AI & ML interests

None yet

Recent Activity

Organizations

Hugging Face's profile picture ๐ŸงจDiffusers's profile picture Hugging Face Internal Testing Organization's profile picture Huggingface Projects's profile picture Snap Research's profile picture Weizmann Institute of Science's profile picture Editing Images's profile picture leditsplusplus's profile picture Latent Consistency's profile picture Editing Audio's profile picture Women on Hugging Face's profile picture +RAIN film festival's profile picture diffusers-internal-dev's profile picture rnri-inversion's profile picture Snapchat Inc.'s profile picture Latent Explorers's profile picture open/ acc's profile picture RF Inversion's profile picture FlowEdit's profile picture LTX Collaborations's profile picture CRINGE's profile picture Rรฉflexion IA's profile picture IP Composer's profile picture Inference Endpoints Images's profile picture

linoyts's activity

reacted to AdinaY's post with ๐Ÿš€ about 6 hours ago
view post
Post
1252
Matrix Game ๐ŸŽฎ an interactive foundation model for controllable game world generation, released by Skywork AI.

Skywork/Matrix-Game

โœจ 17B with MIT licensed
โœจ Diffusion-based image-to-world video generation via keyboard & mouse input
โœจ GameWorld Score benchmark for Minecraft world models
โœจ Massive Matrix Game Dataset with fine-grained action labels
reacted to merve's post with ๐Ÿ”ฅ about 17 hours ago
view post
Post
2816
VLMS 2025 UPDATE ๐Ÿ”ฅ

We just shipped a blog on everything latest on vision language models, including
๐Ÿค– GUI agents, agentic VLMs, omni models
๐Ÿ“‘ multimodal RAG
โฏ๏ธ video LMs
๐Ÿค๐Ÿป smol models
..and more! https://huggingface.co/blog/vlms-2025
  • 1 reply
ยท
reacted to AdinaY's post with ๐Ÿ˜Ž 6 days ago
view post
Post
3870
ACE-Step ๐ŸŽต a music generation foundation model released by
StepFun & ACEStudio

Model: ACE-Step/ACE-Step-v1-3.5B
Demo: ACE-Step/ACE-Step

โœจ 3.5B, Apache2.0 licensed
โœจ 115ร— faster than LLMs (4-min music in 20s on A100)
โœจ Diffusion + DCAE + linear transformer = speed + coherence
โœจ Supports voice cloning, remixing, lyric editing & more
  • 1 reply
ยท
reacted to RiverZ's post with ๐Ÿค— 7 days ago
view post
Post
6000
๐Ÿ”ฅ We're thrilled to share some exciting news about ICEdit! Currently, ICEdit app ( RiverZ/ICEdit) has soared to the second place on the weekly trend list of Hugging Face Space, just trailing behind Qwen3. What's more, it also holds the second position on the overall space trend list. This achievement wouldn't have been possible without your incredible support and love. A huge thank you to each and every one of youโค!

๐ŸŽ‰ The ICEdit community has been incredibly active, and we've seen a plethora of amazing ComfyUI workflows being shared. For instance, with the help of ComfyUI - nunchaku, you can run ICEdit locally with just 4GB of VRAM. This makes it much more accessible for those with limited hardware resources.

๐ŸŽ‡ If you're interested in the detailed information, please head over to our repository. We highly encourage you to give these workflows a try and explore the creative possibilities that ICEdit offers.

Github Repo: https://github.com/River-Zhang/ICEdit
Hugging Face Space: RiverZ/ICEdit
reacted to nyuuzyou's post with ๐Ÿ”ฅ 7 days ago
view post
Post
3551
๐Ÿ–ผ๏ธ PublicDomainFiles.com Collection - nyuuzyou/publicdomainfiles

Collection of 206,204 Public Domain multimedia files featuring:

- Comprehensive metadata: title, description, creator name, keywords, original page URL, and more.
- Contains various media types including images, clip art, artwork, fonts, videos, and TV shows.
- All content explicitly released into the public domain under the CC0 license.
- Organized in a single train split with 206,204 entries.
posted an update 8 days ago
view post
Post
2514
FramePack is hands down one of the best OS releases in video generation ๐Ÿ™‡๐Ÿปโ€โ™€๏ธ๐Ÿคฏ
โœ… fully open sourced + amazing quality + reduced memory + improved speed
but more even - its gonna facilitate *soooo* many downstream applications
like this version adapted for landscape rotation ๐Ÿ‘‡https://huggingface.co/spaces/tori29umai/FramePack_rotate_landscape
  • 2 replies
ยท
reacted to RiverZ's post with ๐Ÿ”ฅ 9 days ago
view post
Post
2664
๐Ÿš€ Excited to Share Our Latest Work: In-Context Edit: Enabling Instructional Image Editing with In-Context Generation in Large Scale Diffusion Transformer๏ฝž

๐ŸŽจ Daily Paper:
In-Context Edit: Enabling Instructional Image Editing with In-Context Generation in Large Scale Diffusion Transformer (2504.20690)


๐Ÿ”“ Code is now open source!
๐Ÿ”ฅ Huggingface DEMO:
RiverZ/ICEdit

๐ŸŒ Project Website: https://river-zhang.github.io/ICEdit-gh-pages/
๐Ÿ  GitHub Repository: https://github.com/River-Zhang/ICEdit/blob/main/scripts/gradio_demo.py
๐Ÿค— Huggingface:
sanaka87/ICEdit-MoE-LoRA

๐Ÿ“„ arxiv Paper:
In-Context Edit: Enabling Instructional Image Editing with In-Context Generation in Large Scale Diffusion Transformer (2504.20690)


๐Ÿ”ฅ Why itโ€™s cool:
- Achieves high-quality, multi-task image editing.
- Uses only 1% of the training parameters and 0.1% of the training data compared to existing methods โ€” extremely efficient
- Beats several commercial models on background preservation, ID control, and consistency
- Open-source, low-cost, faster, and stronger โ€” think of it as the โ€œDeepSeek of image editingโ€ ๐Ÿ‘€

We also implemented a Gradio demo app, available directly in our GitHub repo! And we made a flashy demo video โ€” happy to send it your way!
reacted to abidlabs's post with โค๏ธ 12 days ago
view post
Post
3914
HOW TO ADD MCP SUPPORT TO ANY ๐Ÿค— SPACE

Gradio now supports MCP! If you want to convert an existing Space, like this one hexgrad/Kokoro-TTS, so that you can use it with Claude Desktop / Cursor / Cline / TinyAgents / or any LLM that supports MCP, here's all you need to do:

1. Duplicate the Space (in the Settings Tab)
2. Upgrade the Gradio sdk_version to 5.28 (in the README.md)
3. Set mcp_server=True in launch()
4. (Optionally) add docstrings to the function so that the LLM knows how to use it, like this:

def generate(text, speed=1):
    """
    Convert text to speech audio.

    Parameters:
        text (str): The input text to be converted to speech.
        speed (float, optional): Playback speed of the generated speech.


That's it! Now your LLM will be able to talk to you ๐Ÿคฏ
reacted to ginipick's post with ๐Ÿ‘ 12 days ago
view post
Post
3159
๐ŸŽจ Renoir Studio: Impressionist Masterpieces Reborn Through AI โœจ

๐ŸŒŸ Experience Renoir's Magical Brushstrokes with AI!

๐Ÿ”— Try it now: ginigen/flux-lora-renoir
๐Ÿ”— Model page: openfree/pierre-auguste-renoir
๐Ÿ”— Collection: openfree/painting-art-ai-681453484ec15ef5978bbeb1

Hello, AI art enthusiasts! ๐Ÿ’–
Today I'm introducing a special model - Pierre-Auguste Renoir Studio. Create your own beautiful artwork in the style of the 19th century French Impressionist master! ๐Ÿ–ผ๏ธ
โœจ Why Renoir's Style?
Renoir is famous for his luminous colors and soft brushstrokes. His works feature:

๐ŸŒž Warm sunshine and dancing light
๐Ÿ‘จโ€๐Ÿ‘ฉโ€๐Ÿ‘งโ€๐Ÿ‘ฆ The beauty of everyday life and joyful moments
๐ŸŒธ Vibrant nature and portraits of beautiful women
๐ŸŽญ Lively Parisian social gatherings and outdoor scenes

๐Ÿ”ฌ Technical Features
This model was developed as a flux-based learning model trained on a curated collection of high-resolution masterpieces from renowned global artists. The LoRA fine-tuning process leveraged exceptional quality open-access imagery released by prestigious institutions including the Art Institute of Chicago. The resulting model demonstrates remarkable capability in capturing the nuanced artistic techniques and stylistic elements across diverse historical art movements! ๐Ÿง ๐Ÿ’ซ
๐Ÿš€ How to Use

Describe your desired scene in the prompt box
Add the "renoir" keyword at the end (this is the trigger keyword!)
Click the 'Generate' button
Enjoy your ideas reborn in Renoir's style!

๐Ÿ’ก Recommended Prompt Examples

"Elegant ladies enjoying a picnic in a sunlit garden, wearing pastel dresses and hats renoir"
"People boating by a riverbank, light reflecting on water, warmth of summer renoir"
"Paris cafe terrace, people chatting over coffee, evening sunset renoir"

๐ŸŒˆ Now It's Your Turn!
#AI#Renoir #ArtificialIntelligence#HuggingFace #FLUX #LoRA
reacted to sanaka87's post with ๐Ÿ”ฅ 12 days ago
view post
Post
2503
๐Ÿš€ Excited to Share Our Latest Work: In-Context Edit: Enabling Instructional Image Editing with In-Context Generation in Large Scale Diffusion Transformer๏ฝž

๐ŸŽจ Daily Paper:
In-Context Edit: Enabling Instructional Image Editing with In-Context Generation in Large Scale Diffusion Transformer (2504.20690)

๐Ÿ”“ Code is now open source!
๐Ÿ”ฅ Huggingface DEMO: RiverZ/ICEdit
๐ŸŒ Project Website: https://river-zhang.github.io/ICEdit-gh-pages/
๐Ÿ  GitHub Repository: https://github.com/River-Zhang/ICEdit/blob/main/scripts/gradio_demo.py
๐Ÿค— Huggingface: sanaka87/ICEdit-MoE-LoRA
๐Ÿ“„ arxiv Paper: In-Context Edit: Enabling Instructional Image Editing with In-Context Generation in Large Scale Diffusion Transformer (2504.20690)

๐Ÿ”ฅ Why itโ€™s cool:
- Achieves high-quality, multi-task image editing.
- Uses only 1% of the training parameters and 0.1% of the training data compared to existing methods โ€” extremely efficient
- Beats several commercial models on background preservation, ID control, and consistency
- Open-source, low-cost, faster, and stronger โ€” think of it as the โ€œDeepSeek of image editingโ€ ๐Ÿ‘€

We also implemented a Gradio demo app, available directly in our GitHub repo! And we made a flashy demo video โ€” happy to send it your way!
  • 1 reply
ยท
reacted to jasoncorkill's post with ๐Ÿš€ 15 days ago
view post
Post
5504
๐Ÿš€ Building Better Evaluations: 32K Image Annotations Now Available

Today, we're releasing an expanded version: 32K images annotated with 3.7M responses from over 300K individuals which was completed in under two weeks using the Rapidata Python API.

Rapidata/text-2-image-Rich-Human-Feedback-32k

A few months ago, we published one of our most liked dataset with 13K images based on the @data-is-better-together 's dataset, following Google's research on "Rich Human Feedback for Text-to-Image Generation" (https://arxiv.org/abs/2312.10240). It collected over 1.5M responses from 150K+ participants.

Rapidata/text-2-image-Rich-Human-Feedback

In the examples below, users highlighted words from prompts that were not correctly depicted in the generated images. Higher word scores indicate more frequent issues. If an image captured the prompt accurately, users could select [No_mistakes].

We're continuing to work on large-scale human feedback and model evaluation. If you're working on related research and need large, high-quality annotations, feel free to get in touch: [email protected].
reacted to AdinaY's post with ๐Ÿ”ฅ 16 days ago
view post
Post
5113
Kimi-Audio ๐Ÿš€๐ŸŽง an OPEN audio foundation model released by Moonshot AI
moonshotai/Kimi-Audio-7B-Instruct
โœจ 7B
โœจ 13M+ hours of pretraining data
โœจ Novel hybrid input architecture
โœจ Universal audio capabilities (ASR, AQA, AAC, SER, SEC/ASC, end-to-end conversation)
reacted to samihalawa's post with ๐Ÿ”ฅ 19 days ago
view post
Post
2416
SkyReels-V2 INFINITE VIDEO๐Ÿ”ฅโ™พ๏ธ๐ŸŽฌ UNLIMITED duration video generation model by Skywork.

> โ€œFinally is here. An Open-Source model that achieves what we all have waiting for: Infinite Length Videos.โ€™โ€™๐Ÿ˜ฎ

Skywork R1V: Pioneering Multimodal Reasoning with Chain-of-Thought (2504.05599)

Model: Skywork/SkyReels-V2-T2V-14B-720P

โœจ 1.3B & 14B
โœจ Generates infinite length videos using Diffusion Forcing with diffusion models + autoregressive methods
reacted to victor's post with ๐Ÿ‘ 20 days ago
view post
Post
3460
DIA TTS is just amazing - please share your funniest gens (here is mine) ๐Ÿ˜‚
nari-labs/Dia-1.6B
reacted to AdinaY's post with ๐Ÿ”ฅ 20 days ago
view post
Post
3499
MAGI-1 ๐Ÿช„ the autoregressive diffusion video model, released by Sand AI

sand-ai/MAGI-1

โœจ 24B with Apache 2.0
โœจ Strong temporal consistency
โœจ Benchmark-topping performance
  • 1 reply
ยท
posted an update 21 days ago
reacted to fdaudens's post with ๐Ÿคฏ about 1 month ago
view post
Post
4122
๐ŸŽจ Designers, meet OmniSVG! This new model helps you create professional vector graphics from text/images, generate editable SVGs from icons to detailed characters, convert rasters to vectors, maintain style consistency with references, and integrate into your workflow.

@OmniSVG
  • 2 replies
ยท
reacted to ajibawa-2023's post with ๐Ÿ”ฅ about 1 month ago
view post
Post
3991
Hi All, I recently released two Audio datasets which are generated using my earlier released dataset: ajibawa-2023/Children-Stories-Collection

First Audio Dataset:https://huggingface.co/datasets/ajibawa-2023/Audio-Children-Stories-Collection-Large has 5600++ stories in .mp3 format.

Second Audio Dataset:https://huggingface.co/datasets/ajibawa-2023/Audio-Children-Stories-Collection has 600 stories in .mp3 format.
ยท
reacted to AdinaY's post with ๐Ÿ”ฅ about 1 month ago
reacted to seawolf2357's post with ๐Ÿ”ฅ about 1 month ago
view post
Post
8259
๐ŸŽจ Ghibli-Style Image Generation with Multilingual Text Integration: FLUX.1 Hugging Face Edition ๐ŸŒโœจ

Hello creators! Today I'm introducing a special image generator that combines the beautiful aesthetics of Studio Ghibli with multilingual text integration! ๐Ÿ˜

seawolf2357/Ghibli-Multilingual-Text-rendering

โœจ Key Features

Ghibli-Style Image Generation - High-quality animation-style images based on FLUX.1
Multilingual Text Rendering - Support for Korean, Japanese, English, and all languages! ๐Ÿ‡ฐ๐Ÿ‡ท๐Ÿ‡ฏ๐Ÿ‡ต๐Ÿ‡ฌ๐Ÿ‡ง
Automatic Image Editing with Simple Prompts - Just input your desired text and you're done!
Two Stylistic Variations Provided - Get two different results from a single prompt
Full Hugging Face Spaces Support - Deploy and share instantly!

๐Ÿš€ How Does It Work?

Enter a prompt describing your desired image (e.g., "a cat sitting by the window")
Input the text you want to add (any language works!)
Select the text position, size, and color
Two different versions are automatically generated!

๐Ÿ’ฏ Advantages of This Model

No Tedious Post-Editing Needed - Text is perfectly integrated during generation
Natural Text Integration - Text automatically adjusts to match the image style
Perfect Multilingual Support - Any language renders beautifully!
User-Friendly Interface - Easily adjust text size, position, and color
One-Click Hugging Face Deployment - Use immediately without complex setup

๐ŸŽญ Use Cases

Creating multilingual greeting cards
Animation-style social media content
Ghibli-inspired posters or banners
Character images with dialogue in various languages
Sharing with the community through Hugging Face Spaces

This project leverages Hugging Face's FLUX.1 model to open new possibilities for seamlessly integrating high-quality Ghibli-style images with multilingual text using just prompts! ๐ŸŒˆ
Try it now and create your own artistic masterpieces! ๐ŸŽจโœจ

#GhibliStyle #MultilingualSupport #AIImageGeneration #TextRendering #FLUX #HuggingFace
ยท