-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
djc
committed
Jul 4, 2024
1 parent
9690598
commit f5617bf
Showing
18 changed files
with
332 additions
and
130 deletions.
There are no files selected for viewing
Large diffs are not rendered by default.
Oops, something went wrong.
Large diffs are not rendered by default.
Oops, something went wrong.
Large diffs are not rendered by default.
Oops, something went wrong.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Large diffs are not rendered by default.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -9,32 +9,32 @@ | |
<lastBuildDate>Fri, 21 Jun 2024 11:00:59 -0400</lastBuildDate> | ||
<atom:link href="https://Kwai-Kolors.github.io/index.xml" rel="self" type="application/rss+xml" /> | ||
<item> | ||
<title>γKolors: Unet is enough for diffusionγ</title> | ||
<title>Kolors: Unet is enough for diffusion (arXiv-2024)</title> | ||
<link>https://Kwai-Kolors.github.io/post/post-2/</link> | ||
<pubDate>Fri, 21 Jun 2024 11:00:59 -0400</pubDate> | ||
<guid>https://Kwai-Kolors.github.io/post/post-2/</guid> | ||
<description>Table of Contents π Model Introduction π Evaluation Performance π₯π₯π₯π₯ π₯ Visualization π οΈ Quick Start π License and Citation π Model Introduction We have open-sourced the Kolors large model, which is a large-scale text-to-image generation model based on latent diffusion. The current open-source model has 2.7 billion parameters. Kolors is trained on billions of text-image pairs and demonstrates significant advantages in visual quality, complex semantic understanding, and even text generation (Chinese and English characters) compared to both open-source and closed-source models.</description> | ||
<description>We have open-sourced the <strong>Kolors</strong> large model, which is a large-scale text-to-image generation model based on latent diffusion. The current open-source model has 2.7 billion parameters. Kolors is trained on billions of text-image pairs and demonstrates significant advantages in visual quality, complex semantic understanding, and even text generation (Chinese and English characters) compared to both open-source and closed-source models. Additionally, Kolors supports both Chinese and English, making it more competitive in understanding Chinese-specific content. The model link is provided in the table below; please click to learn more.</description> | ||
</item> | ||
<item> | ||
<title>γLearning Multi-dimensional Human Preference for Text-to-Image GenerationγοΌCVPR-2024</title> | ||
<title>Learning Multi-dimensional Human Preference for Text-to-Image Generation (CVPR-2024)</title> | ||
<link>https://Kwai-Kolors.github.io/post/post-1/</link> | ||
<pubDate>Thu, 20 Jun 2024 10:58:08 -0400</pubDate> | ||
<guid>https://Kwai-Kolors.github.io/post/post-1/</guid> | ||
<description>Authors Sixian Zhang, Bohan Wang, Junqiang Wu*, Yan Liβ‘, Tingting Gao, Di Zhang, Zhongyuan Wang
Links Paper Code arXiv Abstract Current metrics for text-to-image models typically rely on statistical metrics which inadequately represent the real preference of humans. Although recent works attempt to learn these preferences via human annotated images, they reduce the rich tapestry of human preference to a single overall score. However, the preference results vary when humans evaluate images with different aspects.</description> | ||
<description>To learn the multi-dimensional human preferences, we propose the Multi-dimensional Preference Score (MPS), the first multi-dimensional preference scoring model for the evaluation of text-to-image models. The MPS introduces the preference condition module upon CLIP model to learn these diverse preferences. It is trained based on our Multi-dimensional Human Preference (MHP) Dataset, which comprises 918,315 human preference choices across 4 dimensions on 607,541 images. The images are generated by a wide range of latest text-to-image models. The MPS outperforms existing scoring methods across 3 datasets in 4 dimensions, enabling it a promising metric for evaluating and improving text-to-image generation.</description> | ||
</item> | ||
<item> | ||
<title>About Us</title> | ||
<link>https://Kwai-Kolors.github.io/about/</link> | ||
<pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate> | ||
<guid>https://Kwai-Kolors.github.io/about/</guid> | ||
<description>KOLORS Team is an innovative team focused on the research and development of visual generation and editing algorithms. We are dedicated to advancing the forefront of visual generation technology and applying the latest research findings to practical business scenarios.</description> | ||
<description>Kolors Team is an innovative group focused on the research and development of visual generation and editing algorithms. We are dedicated to advancing the forefront of visual generation technology and applying the latest research findings to practical business scenarios.</description> | ||
</item> | ||
<item> | ||
<title>Contact</title> | ||
<link>https://Kwai-Kolors.github.io/contact/</link> | ||
<pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate> | ||
<guid>https://Kwai-Kolors.github.io/contact/</guid> | ||
<description> If you encounter any issues while using KOLORS, feel free to contact us via the email [email protected].
Your Name Email Address An email address is required. Message </description> | ||
<description> If you encounter any issues while using Kolors, feel free to contact us via email [email protected].
Your Name Email Address An email address is required. Message </description> | ||
</item> | ||
</channel> | ||
</rss> |
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Large diffs are not rendered by default.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,26 +1,26 @@ | ||
<?xml version="1.0" encoding="utf-8" standalone="yes"?> | ||
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"> | ||
<channel> | ||
<title>Articles on Kolors</title> | ||
<title>Researches on Kolors</title> | ||
<link>https://Kwai-Kolors.github.io/post/</link> | ||
<description>Recent content in Articles on Kolors</description> | ||
<description>Recent content in Researches on Kolors</description> | ||
<generator>Hugo</generator> | ||
<language>en</language> | ||
<lastBuildDate>Fri, 21 Jun 2024 11:00:59 -0400</lastBuildDate> | ||
<atom:link href="https://Kwai-Kolors.github.io/post/index.xml" rel="self" type="application/rss+xml" /> | ||
<item> | ||
<title>γKolors: Unet is enough for diffusionγ</title> | ||
<title>Kolors: Unet is enough for diffusion (arXiv-2024)</title> | ||
<link>https://Kwai-Kolors.github.io/post/post-2/</link> | ||
<pubDate>Fri, 21 Jun 2024 11:00:59 -0400</pubDate> | ||
<guid>https://Kwai-Kolors.github.io/post/post-2/</guid> | ||
<description>Table of Contents π Model Introduction π Evaluation Performance π₯π₯π₯π₯ π₯ Visualization π οΈ Quick Start π License and Citation π Model Introduction We have open-sourced the Kolors large model, which is a large-scale text-to-image generation model based on latent diffusion. The current open-source model has 2.7 billion parameters. Kolors is trained on billions of text-image pairs and demonstrates significant advantages in visual quality, complex semantic understanding, and even text generation (Chinese and English characters) compared to both open-source and closed-source models.</description> | ||
<description>We have open-sourced the <strong>Kolors</strong> large model, which is a large-scale text-to-image generation model based on latent diffusion. The current open-source model has 2.7 billion parameters. Kolors is trained on billions of text-image pairs and demonstrates significant advantages in visual quality, complex semantic understanding, and even text generation (Chinese and English characters) compared to both open-source and closed-source models. Additionally, Kolors supports both Chinese and English, making it more competitive in understanding Chinese-specific content. The model link is provided in the table below; please click to learn more.</description> | ||
</item> | ||
<item> | ||
<title>γLearning Multi-dimensional Human Preference for Text-to-Image GenerationγοΌCVPR-2024</title> | ||
<title>Learning Multi-dimensional Human Preference for Text-to-Image Generation (CVPR-2024)</title> | ||
<link>https://Kwai-Kolors.github.io/post/post-1/</link> | ||
<pubDate>Thu, 20 Jun 2024 10:58:08 -0400</pubDate> | ||
<guid>https://Kwai-Kolors.github.io/post/post-1/</guid> | ||
<description>Authors Sixian Zhang, Bohan Wang, Junqiang Wu*, Yan Liβ‘, Tingting Gao, Di Zhang, Zhongyuan Wang
Links Paper Code arXiv Abstract Current metrics for text-to-image models typically rely on statistical metrics which inadequately represent the real preference of humans. Although recent works attempt to learn these preferences via human annotated images, they reduce the rich tapestry of human preference to a single overall score. However, the preference results vary when humans evaluate images with different aspects.</description> | ||
<description>To learn the multi-dimensional human preferences, we propose the Multi-dimensional Preference Score (MPS), the first multi-dimensional preference scoring model for the evaluation of text-to-image models. The MPS introduces the preference condition module upon CLIP model to learn these diverse preferences. It is trained based on our Multi-dimensional Human Preference (MHP) Dataset, which comprises 918,315 human preference choices across 4 dimensions on 607,541 images. The images are generated by a wide range of latest text-to-image models. The MPS outperforms existing scoring methods across 3 datasets in 4 dimensions, enabling it a promising metric for evaluating and improving text-to-image generation.</description> | ||
</item> | ||
</channel> | ||
</rss> |
Large diffs are not rendered by default.
Oops, something went wrong.
Large diffs are not rendered by default.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Large diffs are not rendered by default.
Oops, something went wrong.
Large diffs are not rendered by default.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Large diffs are not rendered by default.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters