skip to main content
10.1145/3613904.3642185acmconferencesArticle/Chapter ViewFull TextPublication PageschiConference Proceedingsconference-collections
research-article
Free Access
Artifacts Available / v1.1

TypeDance: Creating Semantic Typographic Logos from Image through Personalized Generation

Authors Info & Claims
Published:11 May 2024Publication History

Abstract

Semantic typographic logos harmoniously blend typeface and imagery to represent semantic concepts while maintaining legibility. Conventional methods using spatial composition and shape substitution are hindered by the conflicting requirement for achieving seamless spatial fusion between geometrically dissimilar typefaces and semantics. While recent advances made AI generation of semantic typography possible, the end-to-end approaches exclude designer involvement and disregard personalized design. This paper presents TypeDance, an AI-assisted tool incorporating design rationales with the generative model for personalized semantic typographic logo design. It leverages combinable design priors extracted from uploaded image exemplars and supports type-imagery mapping at various structural granularity, achieving diverse aesthetic designs with flexible control. Additionally, we instantiate a comprehensive design workflow in TypeDance, including ideation, selection, generation, evaluation, and iteration. A two-task user evaluation, including imitation and creation, confirmed the usability of TypeDance in design across different usage scenarios.

Figure 1:

Figure 1: TypeDance is an authoring tool for creating semantic typographic logos with a flexible and personalized design. By distilling the design principles, it instantiates the design workflow and empowers creators to (a) ideate with an interpretable AI agent, (b) select the typeface at different granularity and imagery from images, (c) generate through blending the selected typeface with targeted imagery, then the creator can (d) evaluate and (e) iterate the position of generated result in the type-imagery spectrum.

Skip 1INTRODUCTION Section

1 INTRODUCTION

Semantic typography is the art of blending typeface and imagery, where the typeface is conceptualized as a visual illustration of semantic representation with high clarity and legibility [21, 23, 47]. One notable application is the semantic typographic logo, which symbolizes a unique identity in a concise yet informative manner. Due to its expressiveness and memorability [7], semantic typographic logo has been widely used as visual signatures for individuals [28], brand logos with commercial values [15, 20], and symbols for significant events and city promotions [3, 43].

However, crafting a semantic typographic logo presents a formidable challenge, requiring seamless blending of typeface and imagery while preserving readability. Experienced designers often rely on professional software like Adobe Illustrator to manually adjust the outline of the typeface to incorporate specific imagery, which is a time-consuming and error-prone process. They often experiment with different strokes or letters of typeface and various imageries to find a visually appealing and memorable representation, intensifying the lengthy process. This requires creative thinking, practical skills, and the ability to persist through continuous trial and error. In addition, the unique identity of a logo necessitates a high level of customization and personalization in the design process.

There are two main challenges in extensive relevant research: blending technique and intent-aware authoring. Existing authoring tools leverage various blending techniques to create other types of designs, e.g., graphical icons. As shown in Fig. 3, One typical technique aims to spatially composite existing materials [13, 66]. Another technique uses shape substitution to achieve a more spatial merge, but it heavily depends on the shape similarity between the objects being blended [8, 9]. Although some computational design techniques support incorporating imagery into typefaces, they are constrained by the ability to change specific parts of typefaces [4, 23, 48]. Advancements in text-to-image generative models [42, 65] have made it possible to generate semantic typography automatically, but it poses another challenge about intent-aware authoring. Given the myriad details such as specific visual presentation (e.g., semantics, color, and shape) in a logo design, text prompts may not be able to represent these intents.

This research aims to gain insights into the design space and workflow involved in creating semantic typographic logos, and then instantiate these design principles to create an AI-assisted tool that facilitates personalized generation. Through analysis of a curated corpus, a systematic design space was identified, focusing on typeface granularity (i.e., stroke-, letter-, and multi-letter-level) and type-imagery mapping (i.e., one-to-one, one-to-many, and many-to-one mappings). Additionally, interviews were conducted with three experts to gather insights into the challenges and concerns regarding AI collaboration in the design process. The findings highlighted the opportunity to simplify the cumbersome blending process and identify a straightforward material for effectively communicating design intentions to generative models.

We propose TypeDance, an authoring tool that empowers both novices and designers with a robust blending technique to create semantic typographic logos from user-customized images. Delighted from “An Image is worth a thousand words” [17, 44] and “Everything you see can be design material” [14, 64], we allow creators to express their design intentions by highlighting the visual representation in their own images. Meanwhile, multiple design inspirations are extracted from the image references for personalizing design. Additionally, we introduce a novel blending technique based on diffusion models that support blending imagery with typeface at all levels of granularity. To guarantee the legibility of both typeface and imagery, we harness a vision-language model to assist creators in pinpointing the position of the generated output within the type-imagery spectrum. We also enable them to edit and refine the output, e.g., making it resemble the typeface “D” more or adopting typewriter-like imagery, as illustrated in Fig. 1. To assess the utility of TypeDance, we conduct the baseline comparison and user study with nine novices and nine designers. Extensive cases and user feedback have revealed the expressiveness of TypeDance in generating diverse semantic typographic logos across different scenarios. In summary, we made the following contributions:

(1)

A formative study that identifies generalizable design patterns and simulatable design workflow.

(2)

An intent-aware input based on user-personalized image that goes beyond ambiguous text prompt, providing a detailed visual description of the desired logo design.

(3)

A blending technique that incorporates imagery with all levels of typeface granularity.

(4)

An authoring tool that integrates a comprehensive workflow, empowering creators to ideate, select, generate, evaluate, and iterate their designs.

(5)

A two-task evaluation and a baseline comparison that show the usability of TypeDance across different usage scenarios.

Skip 2RELATED WORK Section

2 RELATED WORK

2.1 Semantic Typographic Logo Design

Semantic typographic logos are harmonious integration of typeface and imagery, where the imagery is visually illustrated by typeface [23, 43, 48]. Compared with plain wordmark [50, 52] and pictorial logo [22, 30], semantic typographic logo allows a more cohesive approach to encode both word and graphic content and enhance the association between them. The capacity to embody rich symbolism and expressiveness has led to increasing adoption of semantic typographic logos across various scenarios, such as cultural promotion [3], commercial brand [15] and personal identity [28]. Extensive research has explored how typefaces can be designed to reinforce semantic meaning at varying levels of granularity. Some studies subdivide typeface into a series skeletal strokes with user-guided [38] and automatic segmentation [4], and then apply structural stylization to each stroke and junction separately. In contrast, recent studies [23, 48, 59] have shifted their focus from stroke-level stylization to individual letter stylization using predefined templates. For instance, Tendulkar et al. [48] replaced letters with clipart icons relevant to the imagery and visually resembling the corresponding letter. Another approach, as demonstrated by Xu et al. [55], involves compressing the multi-letter and arranging them into a predetermined semantic shape. This approach has been further enhanced by Zou et al. [68], who proposed an automatic framework that supports the placement, packing, and deformation of compact calligrams.

While prior research extensively investigated the semantic typographic logo across different typeface design granularities, two key issues persist: 1) these models are constructed for typefaces with specific granularity, limiting their applicability, and 2) little is known regarding the mapping relationship between typeface and imagery. These works typically employ a simple approach where one typeface is paired with one specific imagery. To explore the design space, we collect a real-world corpus, analyze typeface granularity and type-imagery mapping, and instantiate these design principles in TypeDance. Then we propose a unified framework based on diffusion model to support flexible blending between imagery and typefaces at different granularities.

2.2 Generative Model for Computational Design

Computational design has garnered considerable attention in the field of generative techniques. Recently, there have been advancements in aligning semantic meaning between image and text pairs, making natural language a valuable tool that bridges the gap between humans and creativity [29, 39]. Numerous studies have exploited such semantic consistency to retrieve relevant images from the corpus using natural language statements, which can be used as design materials to generate new designs [12, 66]. While previous studies relied on retrieving from limited corpus and predefined templates, more recent research has proposed text-to-image diffusion models [40, 47] that surpass mainstream GAN models[18] and autoregressive models[41]. However, this plain text-guided generation relies heavily on well-designed prompts, leading to unstable results devoid of user control. To address this issue and enhance user customization, recent advancements have introduced image-based conditions for achieving controllable manipulations, including depthmap [42] and edgemap [65]. Some generative models focusing on font stylization only support the letter-level generation [23] and require collecting images containing the specific imagery for fine-tuning the model [47].

While prior works have demonstrated incredible generative ability in creating complex structures and meaningful semantics, ensuring the readability of both the typeface and the imagery remains a daunting task. In particular, the text condition lacks sufficient restrictions to capture all user intentions, while the image condition is overly rigid and cannot accommodate the inclusion of additional information. To tackle this challenge, Mou et al. [34] proposed an approach that combines multiple conditions to improve controllability. Similarly, Vistylist [45] disentangles the design space, enabling the generation with combined user-intended design factors. TypeDance builds upon these previous research efforts by providing several design priors that allude to the characteristics of semantic typographic logos. These design priors extracted from user-provided images serve as guidance for users to select and incorporate into their designs. With support for both text and image conditions, TypeDance empowers users with flexible control, enabling personalized and distinctive design outcomes.

Figure 2:

Figure 2: A general workflow for semantic typographic logo design is outlined from the expert interview, with the main challenges in the workflow and concerns of generative AI labeled on corresponding stages. Based on the workflow, challenges, and concerns, the design consideration of Typedance is solidified.

2.3 Graphic Design Authoring Tool

Significant works have developed tools to facilitate graphic design, which can be broadly divided into two primary categories: ideation and creation tools. In the domain of ideation, several research studies [24, 26, 56] have proposed interfaces aimed at inspiring ideas and facilitating the exploration of design materials. For example, MetaMap [24] employed a mindmap-like structure encompassing three design dimensions to stimulate users and encourage them to generate a wide range of unique and varied ideas. Regarding the creation process, as Xiao et al. [54] identified, mainstream works follow a two-stage pipeline, which involves retrieving examples and adapting them as design material [64] and style transfer reference [45]. More recently, researchers sought to blend approaches to create a novel design based on existing design materials. During the process, spatially compositing semantically related icons to generate a compound design in a resourceful manner is adopted by some researchers [13, 66]. Similarly, Zhang et al. [63] demonstrated that compositing coherent imagery elements can create an ornamental typeface with wide conceptual coverage. On the other hand, Chilton et al. [8, 9] further explored the potential of blending through similar shape substitution. For instance, they showed that the “Starbucks logo” can replace the position of the “sun” as both have a circular shape.

However, spatial composition and shape substitution techniques encounter challenges when dealing with the complexity of semantic typographic logos, in which typeface and imagery need to be spatially fused as a whole despite the absence of shape similarity. In this work, Typedance utilizes diffusion models to incorporate imagery detail while preserving the salient representation of the typeface, enabling a more natural blend. Additionally, Typedance integrates both ideation and creation functions. To ensure the readability of both the typeface and the imagery in semantic typographic logos, an evaluation component is further incorporated, enhancing the faithfulness of the design process.

Skip 3FORMATIVE STUDY Section

3 FORMATIVE STUDY

To instantiate the real design principles in TypeDance, we extracted simulatable design workflow from semi-structured interviews and generalizable design patterns from a corpus analysis.

Participants. We conducted semi-structured interviews with three experts: a design professor leading logo design teams for international conferences and city identities (E1), a brand designer with over 11 years of corporate and startup logo design experience (E2), and a logo designer who has received several renowned design prizes (E3). All three experts have extensive experience in semantic typography design.

Procedure. Each individual interview, lasting one to one and a half hours, began with a presentation of the interviewee’s work from social media. We then delved into their interpretation and detailed explanation of the design process. Finally, we posed questions about key steps in creating a semantic typographic logo, the most challenging step, and expectations and concerns regarding generative models.

3.1 General Workflow and Challenges

3.1.1 General Design Workflow.

Semantic typography design is a creative process that involves generating innovative ideas and implementing them. As Fig. 2 shows, The general workflow typically consists of five steps.

Ideation. To begin with, designers are often given fixed text, such as a brand name. Based on the usage scenarios, they need to come up with innovative ideas regarding potential imagery.

Selection. Once designers confirm a design idea, they prepare design materials with two crucial aspects: 1) the part of the typeface structure and 2) the specific visual representation of imagery. The choice of typeface evolves through multiple attempts with different granularities, such as a single stroke or the entire typeface. All experts mentioned that they obtained inspiration for visual representation from images, whether provided by customers, collected from design-sharing communities, or from their own life gallery. Images serve as a source of other inspiration as well. As E1 noted, “I usually discover new inspiration in pictures, such as color palette.

Generation. Designers start by simplifying the visual representation into a basic shape that corresponds to the typeface’s skeleton in the sketch. They then adjust the typeface outline using professional software like Adobe Illustrator, seamlessly integrating it with the chosen imagery.

Evaluation. Upon completing a design, designers will assess the legibility of both the typeface and imagery incorporated in their work. It often hinges on external validation from individuals other than the designers themselves.

Iteration. Iteration is conducted throughout the design process. Designers conduct multiple experiments and refinements at each step to reach potential outcomes. Refinement continues until a finalized design is achieved.

3.1.2 Challenges in Workflow.

Figure 3:

Figure 3: (a) The comparison of blending technique between prior works and TypeDance. (b) Some semantic typographic logo examples from the corpus, each labeled with the corresponding design pattern, including Typeface Granularity and Type-Imagery Mapping.

Depletion of ideas during brainstorming. Experts widely regard the birth of good inspiration as a combination of imagination and serendipity. E2 emphasizes the importance of delving into the story behind a brand and then incorporating it into logo design. This involves gathering background knowledge and discovering imagery that resonates with human perception. Diverse plans for alternatives are essential for iteration.

Clarification of specific visual representation of imagery. Designers frequently explore diverse parts of the typeface while experimenting with various imagery options to attain the most optimal and visually appealing outcomes. However, a single image can be depicted in various visual presentations, as illustrated by the diverse robots in Fig. 2, posing a challenge in specifying a particular one. Additionally, the compatibility of the imagery with the typeface can complicate the selection process.

Tedious and laborious manipulation in blending typeface and imagery. Crafting a logo design from a draft involves conceptualizing and professionally implementing the blending of significantly different imagery and typeface. As mentioned by E3, “sometimes it is challenging when my client insists on having both the letter ‘M’ and a standing cat from the photo she gave me – finding similarities in their shapes is hard.” For the implementation, despite relying on professional software like CorelDRAW and Adobe Illustrator, the blending process remains manual, consuming a substantial amount of time and effort.

Challenges in evaluating legibility. A successful semantic typographic logo should not confuse the public about its typeface or meaning. However, evaluating the legibility often falls to designers’ subjective judgment rather than being based on the perception of the general public. This poses a challenge in achieving a fair evaluation.

3.2 Concerns in Generative Model Involvement

With the rise of generative models like Midjourney, many designers began to embrace AI for design assistance. The following are two main concerns about AI involvement in their design process.

Lack of controllability of text-to-image generative model. The most widely used method for controlling generative models is through textual prompts. Though various tools can assist in designing prompts, like PromptBase1 and PromptHero2, these tools emphasize imitating certain styles and lacking precise control of specific shape and layout. When dealing with highly personalized imagery, the time and effort invested in crafting a prompt experiences a significant surge. As E1 notes, “Each time I intricately design a prompt in anticipation of the generated result, it prompts me to open Illustrator again.

Lack of refinement and editability of generative result. Lack of editability in the results appears to be a common issue of generated models. While users may be generally satisfied with the overall outcome, there can be instances where certain details may not meet their expectations (e.g., dislike colors, redundant objects). One approach to address this is to regenerate the entire image, which results in losing the current design.

3.3 Design Space of Semantic Typography Work

To further identify design patterns shaping semantic typographic logos, we collect and analyze a corpus of 427 real-world examples3. To ensure the diversity of sources, we include examples from prior research [23, 63], reputable design communities4, and influential design shared on social media. The keywords we mainly use for search are “topographic logo,” “semantic topography,” and “word as image”. Focusing on logogram (Chinese (97), Japanese (34), Korea (30)) and alphabet language (English (229), French(20), Russian(17)), we filter search engine results for each language based on “popular” and “new” criteria, considering both widespread acknowledgment and timeliness. As Fig. 3 shows, our corpus analysis reveals two critical aspects of design patterns: (1) typeface granularity and (2) type-imagery mapping. Besides language types, typeface granularity, and type-imagery mapping, we also encode logo usage scenarios. Two authors served as coders to ensure objectivity with a double check.

3.3.1 Typeface Granularity.

While various languages have their own unique symbols, they conform to a shared structural granularity. This hierarchy, arranged from local to global, encompasses stroke, letter, and multi-letter, respectively.

Stroke-level. It often employs stroke decomposition, where a single stroke or a group of strokes is associated with imagery (123/427 examples). As illustrated in Fig. 3 (b1), the spout of a teapot aligns with the original curve in the letter “g”, enhancing semantic expression while preserving typeface integrity. As the smallest unit of typeface, stroke-level blend can be implemented multiple times within a typeface to enrich visual representation.

Letter-level. Individual letters are commonly used in our collected corpus (189/427 examples). Rather than conducting a letter-level blend to every single letter, certain examples tend to focus on partially representative letters within a word, particularly the first letter in Fig. 3 (b6). To emphasize the imagery, they employ techniques such as scaling, elongation, and rotation on the letters.

Multi-letter-level. Blending imagery with multiple letters or entire words is the multi-letter-level blend (115/427 examples). It regards the typeface as a cohesive unit and spatially arranges the letter in a proper position. According to Fig. 3 (b5), the letters are rearranged and distorted to create the recognizable silhouette of a dog.

3.3.2 Type-Imagery Mapping.

We observe a complex linkage between typeface and imagery. Prior works simplify such linkage by adopting a single mapping strategy, where one letter is associated with specific imagery. To approximate TypeDance to the design practice of semantic typographic logo, we manually encode the corpus and identify three typical mapping patterns: one-to-one, one-to-many, many-to-one. With a comprehensive understanding of how typeface interact with imagery, we can better instantiate the design principles to empower the creation process.

One-to-One Mapping. One logo corresponds to one imagery commonly observed in the corpus (294/427 examples). It preserves typeface structures using partial strokes or letters to represent the same imagery. For instance, in Fig. 3 (b1), the letter “g” incorporates the imagery of a teapot spout. Additionally, we observed that logos with one-to-one mapping often employ repetitive imagery with a consistent style within a particular typeface.

One-to-Many Mapping. The semantic typographic logo in this portion will distribute multiple imageries in the typeface involved in the design (14/427 examples). This mapping type supports rich imagery coverage within a compact space, where the semantic concepts usually share the same theme. Fig. 3 (b4) integrates both spoon and fork into a letter “m”, underscoring the theme of the meal.

Many-to-One Mapping. Another aspect involves integrating multiple letters in typeface into a single imagery (119/427 examples), typically achieved by combining entire words to convey a complex visual representation of meaning, see Fig. 3 (b3, b4). This creative approach can be traced back to Giuseppe Arcimboldo, who skillfully merged various elements and shapes to create cohesive portraits and figures [32]. The many-to-one mapping enhances the overall unity and deepens the expression of semantic meaning.

3.3.3 Summary.

Through corpus analysis, we uncover different typeface granularity and various type-imagery mapping. The combination of these design patterns presents an opportunity to convey rich visual representations. Both logograms and alphabet languages adhere to these patterns but exhibit distinct preferences. The complex typeface structure of logograms and their inherent pictorial origins result in more elaborate imagery combinations. For instance, compared to the French word in Fig. 3 (b5), the Chinese words in Fig. 3 (b6) achieve a blending with traditional landscape paintings without spatially rearranging the typefaces. The correlation between granularity and whether the logo consists of letters (alphabet) or characters (logograms) is evident in the collected corpus. As granularity increases, logos with letters show the following distribution: 9.40% (25/266), 55.64% (148/266). For logos with characters, the distribution is 60.87% (98/161), 25.47% (41/161), 13.66% (22/161). Given the complex structure inherent in characters, local incorporation with imagery is beneficial to maintain legibility, whereas letters with complex structures allow for more global transformation.

Additionally, we note distinctions in real-world logo design compared to the formal definition of semantic typography. In real-world logo design, blending is not uniformly applied to all typefaces. Instead, emphasis is placed on letters in the initial position or those closely related semantically to the imagery (171/427 examples). Moreover, in significant designs, the typeface often has a less direct semantic relationship with the incorporated imagery (203/427 examples). This observation aligns with insights from an interview with E2, who remarked, “ The selection of imagery depends on the brand’s story, and the logo’s meaning is for users to associate the imagery directly with the brand.

Skip 4DESIGN CONSIDERATION Section

4 DESIGN CONSIDERATION

The expert interview reveals that personalizing a semantic typographic logo relies on blending specific typefaces (e.g., at different granularity) and imagery (e.g., concrete visual representation), further identified through corpus analysis. Challenges in user workflow and concern about generative AI highlight using easily accessible images for effective personalization, eliminating the need for intricate text prompts that may not fully capture user intentions. Guided by Shneiderman’s design principle [46] of “Design with low thresholds, high ceilings, and wide walls,” our goal is to develop a tool that enables novices to create with accessible materials and interactions ( D1 ), automates complex blend manipulation for professionals ( D2 ), and integrates essential functionalities for a streamlined design process ( D3 ). The roadmap we derive the design considerations is illustrated in Fig. 2. Below, we present the set of design considerations:

D1.

Intent-aware design material and interaction. We aim to support flexible material selection, allowing the easy switch between typefaces at different granularities and the selection of imagery from visual representations in a user-customized image.

D2.

Facilitate the professional generation process. We aim to propose an automatic blending approach that supports typefaces at all levels of granularity, ensuring harmonious and diverse designs.

D3.

Provide necessary functionalities to support a comprehensive workflow. In the pre-generation stage, we will incorporate an ideation module for brainstorming. In the post-generation stage, an evaluation and iteration module will be added to identify, edit, and refine the generated result within the type-imagery spectrum.

Skip 5TYPEDANCE Section

5 TYPEDANCE

Based on the design rationales and the opportunity to address the challenges and concerns in the design workflow, we introduce TypeDance, an authoring system that facilitates personalized generation for semantic typographic logos. TypeDance comprises five essential components, which closely correspond to the pre-generation, generation, and post-generation stages. In the pre-generation stage, the ideation component communicates with the user to gather comprehensible imageries (D1). The selection component allows the user to choose specific design materials, including typeface at various granularities and imagery with particular visual representations (D2). The generation component blends these design materials, utilizing a series of combinable design priors (D3). After the generation, the evaluation component enables the user to assess the current result’s position in the type-imagery spectrum (D4). The iteration component empowers user to refine their design by adjusting the type-imagery spectrum and editing each individual element (D5).

5.1 Ideation

Thanks to the impressive language understanding and reasoning capabilities of large language models, we can now collaborate with a knowledgeable brain through text. TypeDance takes advantage of Instruct GPT-3 (davinci-002) [36] with Chain-of-Thought prompting [53] to generate relevant imagery based on given texts. To enhance user understanding, the prompting strategy includes the requirement to accompany the imagery with explanations in terms of visual design. This ensures that the answers provided are more interpretable and informative. For example, when the user provides the keyword “Hawaii,” TypeDance generates concrete imagery such as “Aloha Shirt,” “Hula Dancer,” and “Palm tree.” Along with these imagery words, TypeDance also provides explanations like “Symbolizes the vibrant culture and traditional dance form of Hawaii” for the “Hula Dancer”. By making this small change, TypeDance provides interpretable explanations and offers users additional background knowledge to enhance their understanding.

5.2 Selection

It aims to prepare design materials blended in the following generation. It encompasses two fundamental components: selecting typeface It at various granularities and imagery Ii with particular visual representations. We achieve the fine-grained high-fidelity segmentation based on Segment Anything Model [25], which offers user-friendly visual prompts as input, including box and point. As illustrated in Fig. 4, the different characteristics inherent to typefaces and imagery give rise to varying interactions that aim to align with the design rationales.

Typeface Selection. Given the text, TypeDance allows creators to select typefaces It at different granularity, as identified in our formative study, to achieve a more fine-grained and flexible approach to complex design practices. Instead of being limited to using complete strokes, TypeDance enables creators to select partial regions of a single stroke. To achieve this, As Fig. 4 shows, we implemented the drag-select interaction for creators to encompass specific parts of the typeface they need within a designated box. Compared to the click interaction, the drag-select offers a more explicit way to reveal user intention, which is free from rule-based segmentation that is restricted by a set of predetermined strokes, thus supporting more fine-grained selection. Moreover, TypeDance offers a combination selection by which creators can select strokes located far apart, as in the case shown in Fig. 4.

Imagery Selection. Regarding imagery selection, we employ semantic segmentation to extract the visual representation creators require from a cluttered background. However, unlike typeface selection, the nature of imagery selection is more conducive to clicking rather than drag-selection. As depicted in Fig. 4, drag-selection can inadvertently encompass other objects the creator may not need while selecting their desired object. To address this issue, we have implemented a solution where creators can click on individual objects separately, mitigating the problem of unintentional object coverage. Similarly to typeface selection, TypeDance also supports combination selection, allowing creators to choose multiple objects within the image.

Figure 4:

Figure 4: The Selection and Generation component in the workflow of TypeDance. The selection component offers two types of interaction to allow creators to flexibly select typeface It at different granularity and imagery It with specific visual representation. These design materials will be injected into the diffusion model in the generation component with an optional user prompt It. The discrimination is conducted to ensure the generated result can meet three-dimensional user intent, including It, Ii, and Tp.

5.3 Generation

5.3.1 Input Generation.

This section describes the three inputs required for the generation process. The first input is the selected typeface It, which serves as the origin image for the diffusion model. The second input is the optional user’s prompt Tp, which allows them to explicitly express their intent, such as the specific style they desire. The third input consists of the design factors extracted from the selected image Ii.

Semantics. Textual prompt is an accessible and intuitive medium for creators to instruct AI, which also offers a way to incorporate imagery into the generation process. However, it is laborious to describe a significant amount of information within the constraints of a limited prompt length. TypeDance solves this problem by automatically extracting the description of the selected imagery. Describing the selected imagery involves a text inversion process encompassing multiple concrete semantics dimensions. One of the prominent semantics is the general visual understanding of a scene. For instance, in Fig. 4, the description of the scene is “a yellow vase with pink flowers.” We capture this explicit visual information (object, layout, etc.) using BLIP [29], a Vision-Language model that excels in image captioning tasks. Moreover, the style of imagery, especially when it comes to illustrations or paintings, can greatly influence its representation and serve as a common source of inspiration for creators. The style of the case in Fig. 4 is “still life photo studio in style of simplified realism.” Such a specific style is derived from retrieving relevant descriptions with high similarity in a huge prompt database. Therefore, the complete semantics of the imagery include the scene and style. To enhance interface scalability, we extract keywords from the detailed semantics. Creators can still access the complete version by hovering over the keywords.

Color. Considering color is one of the most important design inspiration sources [60, 61], TypeDance utilizes kNN clustering [16] to extract five primary colors from the selected imagery. These color specifications are then applied in the subsequent generation process. In order to preserve the semantic colorization relation, the extracted colors are transformed into a 2D palette that includes spatial information. This ensures that the generated output maintains a meaningful and coherent color composition.

Shape. The shape of the typeface can take an aesthetic distortion to incorporate rich imagery, as demonstrated in our formative study. To achieve this, we first leveraged edge detection to recognize the contour of selected imagery. Then, we sample 20 equidistant points along the contour. These points are used to deform the outline of the typeface iteratively, using generalized Barycentric coordinates [33]. The deformation occurs in the vector space, resulting in a modified shape that depicts coarse imagery and facilitates guided generation.

These design factors are applied independently during the generation process. Creators have the flexibility to combine these factors according to their specific needs, allowing for the creation of diverse and personalized designs.

5.3.2 Output Discrimination.

To ensure that the generated result aligns with the creators’ intent, TypeDance employs a strategy that filters good results based on three scores. As illustrated in Fig. 4, we aim for the generated result Ig to achieve a relatively balanced score in the triangles composed of typeface, imagery, and the optional user prompt. The typeface score s1 is determined by comparing the saliency maps of the selected typeface and the generated result. Saliency maps are grayscale images that highlight visually salient objects in an image while neglecting other redundant information. We extract the saliency maps for the typeface and the generated result and then compare their similarity pixel-wise. The imagery score s2 is derived from the cosine similarity between the image embeddings of the input image Ii and the generated result Ig. Similarly, we obtain the prompt score s3 by computing the cosine similarity between the image embedding of the generated result Ig and the text embedding of the user prompt Tp. We use the pre-trained CLIP model to obtain the image and text embeddings because of its aligned multi-modal space. We denote si = {si1, si2, si3}, where i represents the i-th result at one round of generation. To filter the results that mostly align with the creators’ intent, we use a multi-objective function that maximizes the sum of the scores and minimizes the variance between them. The function is defined as follows: \(\begin{eqnarray*} \max _{s} f(\mathbf {S}) = \sum _{j=1}^{3} s_{ij} - \lambda \cdot \sigma (\mathbf {S}), \end{eqnarray*} \) where S is the score set of all generated results, and σ(S) calculates the variance of the scores. The λ is a weighting factor used to balance the total score and variance, which is empirically set as 0.5. Based on this criteria, TypeDance displays the top 1 result on the interface each round and regenerates to obtain a total of four results.

5.4 Evaluation

Once the design is generated, evaluating a design’s compliance with recognized visual design principles is crucial for its completion [9, 37]. Semantic typographic logos require a balance between the typeface’s readability and the imagery’s expressiveness, presenting a challenging tradeoff. Designers often rely on the feedback from participants to validate their designs. Typdance offers a data-driven instant assessment before gathering participants’ reactions.

To assist users in determining the position of their current work on the type-semantic spectrum, TypeDance utilizes a pre-trained CLIP model [39] that provides objective judgment supported by data. By distilling knowledge from a vast dataset of 400 million image-text pairs of CLIP model, TypeDance can quantify the similarity between typeface and imagery on a scale ranging from [0, 1], with the sum equal to 1. To enhance intuitive understanding, TypeDance translates the neutral points from 0.5 to 0 and normalizes the distance between the similarity and the neutral point from 0.5 to 1. Users can determine the degree of divergence from the neutral point for their currently generated result. A value of 0 signifies neutrality, indicating that the generated result favors both the typeface and the imagery. Conversely, higher values indicate a greater degree of divergence towards either the typeface or the imagery, depending on the specific direction.

5.5 Iteration

Although iteration is throughout the process, we identified three main iteration patterns.

5.5.1 Regeneration for New Design.

Previous tools often allow creators to delete unsatisfactory results, but few aim to improve the regenerated performance to meet creators’ intent. To address this challenge, TypeDance utilizes both implicit and explicit human feedback to infer user preferences. Inspired by FABRIC [49], we leverage users’ reactions toward generated results as a clue to user preference. By analyzing positive feedback from preserved results and negative feedback from deletions, the generative model in TypeDance dynamically adjusts the weights in the self-attention layer. This iterative incorporation of human feedback allows for the refinement of the generative model over time. Additionally, TypeDance provides users with more explicit ways to make adjustments, including a textual prompt and a slider that enables users to control the balance between typeface and imagery.

5.5.2 Refinement in Type-imagery Spectrum.

TypeDance provides a refined approach to iteratively refine their designs along the type-imagery spectrum. In addition to the quantitative metric identified in the evaluation component, TypeDance allows creators to make precise adjustments using the same slider. As shown in Fig. 5 (c), we divide the distance between the neutral point and the typeface and imagery into 20 equal intervals, each representing 0.05. These small interpolations preserve the overall structure and allow for incremental adjustments. By dragging the slider, creators can set their desired value point between typeface and imagery, achieving a balanced aesthetic.

Specifically, to prioritize imagery, we begin with the current design as the initial image and inject it into the diffusion model with the strength set to the desired value point. This process pushes the generated image toward the imagery end, resulting in a more semantically rich design. Conversely, to emphasize the typeface, we utilize the saliency map of the typeface in the pixel space to filter out the relevant regions of the image. This modified image is fed into the generative model to ensure a smooth transition towards the typeface end.

Figure 5:

Figure 5: The interface of TypeDance, with a creator engaging in semantic typographic design. (a) In pre-generation, creator brainstorms for ideas and selects typeface and imagery as design materials. (b) During generation, creator sets generation options along with a prompt to personalize the design. (c) In post-generation, the creator evaluates and refines the design in the type-imagery spectrum.

5.5.3 Editability for Elements in the Final Design.

A fine-grained adjustment is required to make nuanced changes to an almost satisfactory result, such as deleting an element in the generated result. In order to achieve this level of editability, we convert the image from pixel to vector space. Hence, creators gain the ability to manipulate each individual element in their design, allowing them to remove, scale, rotate, and change colors as needed.

Skip 6INTERFACE WALKTHROUGH Section

6 INTERFACE WALKTHROUGH

In TypeDance, the components that follow the design workflow are organized in a cohesive U-shaped layout on the interface, with the main canvas at the center, as depicted in Fig. 5. This design aims to facilitate seamless navigation for creators, eliminating the need for constant switching between different components. Following the user workflow, we demonstrate how TypeDance creates visually appealing designs. Alice is a graphic designer who wants to create a series of postcards for the four seasons using imaginative semantic typographic illustrations. Starting her creative journey, Alice begins with the character “,” which means “Spring” in Chinese.

6.1 Pre-generation stage

6.1.1 Seeking design ideas.

To gather design inspiration, Alice first types the “Give me some ideas about spring” in the IDEATION and click the [Brainstom] button. TypeDance generates a list of ideas like “Bird in Nest” and “Blooming Flowers”. She hovers over these ideas, and their corresponding explanations pop up.

6.1.2 Preparing design materials.

Alice types the character “” in the TYPEFACE and selects the font type as “Mincho”. She uses the drag-selection to choose the lower part of the character, represented as “,” and clicks [✓ ] button to confirm her selection. Consequently, the rest of the typeface appears on the central canvas. Next, she opens another tab to search for images using the keyword “Blooming Flowers in Spring”. After selecting an appropriate image, she uploads it to the IMAGERY by clicking to select the window object within the image. Upon doing so, the selected window object is highlighted. To finalize her selection, Alice clicks the [✓ ] button.

6.2 Generation stage

6.2.1 Blending typeface and imagery.

Alice browses the extracted design factors in the GENERATE and selects the [semantic] and [color] options. She leaves the strength at its default setting of 0.75 for her first attempt. To add more imagery of spring, she looks back to the IDEATION and notices the “Bird in Nest”. she proceeds to input the prompt “bird on the window” in the GENERATE and submit the generation. After a brief 15-second wait, the results are presented in the GALLERY .

6.2.2 Regenerating with appropriate strength.

Recognizing the generated result is closer to the typeface, she deletes unwanted results and adjusts the design factor strength to 0.86 using the slider. In the subsequent round, she finds a desirable result and clicks it. The chosen design is then displayed in the central canvas.

6.3 Post-generation stage

6.3.1 Evaluating and Refining the generated result.

To assess its legibility, Alice navigates to the right side of the canvas and clicks the [Evaluation] button in the EVALUATION. The current position of the result is situated on the imagery side of the slider with a value of 0.55. However, aiming to explore positions more aligned with the typeface side, she drags the slider to the left. After several trials, Alice obtains a series of results, as shown in Fig. 5.

6.3.2 Editing and Exporting.

Alice repeats the process to blend a stroke “” with the floral imagery. She converts them into SVG format and modifies the color of the flowers. Finally, Alice exported the design by clicking the [save] button on the top-right corner of the interface.

Skip 7EVALUATION Section

7 EVALUATION

To test the effectiveness of TypeDance, we conducted a baseline comparison and user study. We aim to evaluate the performance of generated results and the usability of TypeDance, and explore how each component could potentially address pain points in their design workflows. Additionally, we delved into the limitations of the tool and identified opportunities for improvement.

7.1 Baseline Comparison

We conducted a comparison with seven alternative methods: Zhang et al. (M1)[63], TReAT (M2)[48], Word as Image (M3)[23], DS-Fusion (M4)[47], Depth2Image (M5)[34], ControlNet (M6)[65], Dalle 3 (M7)[5]. For a comprehensive evaluation, we assessed these methods from both technical and perceptual perspectives, as depicted in Fig 6. Given that most works are not open-source, an overall comparison using the same case was not feasible. Instead, we randomly sampled three cases from each method and utilized TypeDance to recreate them with the same text content and imagery, taking a one-to-one comparison with each method. The full cases are listed in the supplemental material. The perception study involved an online questionnaire with the participation of 50 individuals. We shuffled the appearance sequence of all logos and provided no hints about the original typeface or imagery used, and the scores were recorded on a 5-point Likert scale.

Figure 6:

Figure 6: TypeDance vs. baselines: (a) comparison in technique, design material, and perception, and (b) comparison in case performance.

7.1.1 Technique Difference.

Artistic typography and TReAT operate on a retrieval-based approach, which limits their ability to generalize to cases significantly different from the collected templates. In contrast, other methods adopt a generation-based approach, offering some alleviation to this limitation. However, most of them are restricted to letter-level blending, whereas TypeDance excels by supporting blending at all typeface granularity. Regarding the interaction to specify imagery, the majority of methods rely on text, either by retrieving relevant templates from a corpus or guiding the generative model. Notably, DS-Fusion and TypeDance stand out as they support using images to assign specific visual representations. It’s worth noting, though, that DS-Fusion necessitates users to provide a small image dataset of around 20 images for fine-tuning the model, a process taking approximately 1.5 hours using a desktop with Nvidia GeForce RTX 3090.

7.1.2 Perception Study.

We used two primary metrics to assess the performance of these methods: one focused on the legibility of typeface and imagery, and the other on aesthetics. As illustrated in Fig. 6, TypeDance outperforms other methods in both aesthetics score and legibility of imagery. Word as Image achieves the highest score in the legibility of typeface, but its imagery is comparatively challenging to recognize. Similarly, many methods exhibit this imbalance, excelling in one representation while compromising the other. In contrast, TypeDance maintains a stable performance with commendable aesthetic recognition.

7.2 User Study

7.2.1 Participants.

Both designers and general users (11 females and 7 males, aged 19-34) are invited to obtain different feedback. Nine participants (P1-P9) are novice users interested in semantic typography art without formal design training. The remaining nine are graphic designers (E1-E9) with professional design education with more than three years of experience in semantic typographic logos. All participants have tried AI tools like Midjourney before. They accessed TypeDance through web browsers, utilizing a combination of online and offline modes. As a token of appreciation, each participant received a $30 gift card upon completing the study.

Figure 7:

Figure 7: The first task of evaluation is imitation. In this task, users are required to choose two out of three references and imitate their style. Part of the design outcomes created by participants via TypeDance are shown, with annotations indicating the necessary design materials, including typeface and imagery, and the design patterns in their designs.

7.2.2 Tasks and Procedure.

Drawing inspiration from the design process outlined by Okada et al. [35], which contains two essential stages, namely imitation and creation, we crafted two tasks accordingly to align the natural and progressive design challenges. The comprehensive procedure is outlined as follows:

Briefing (20 minutes). We presented several examples of semantic typographic logos from our collected corpus to introduce the topic. To facilitate a swift transition for participants into their roles as creators, we inquired about their preferred design and asked them to envision the steps required to accomplish such a design. Next, we introduced TypeDance using the example depicted in Fig. 4. Participants were then encouraged to independently explore TypeDance for 5 minutes to gain further familiarity with its functionalities.

Task1: Imitation (25 minutes). Participants are instructed to select two of three design references and replicate the style in their own designs. Reference 1 encourages participants to incorporate imagery representing their passions in the typeface “love” at the letter-level. Reference 2 endeavors to integrate as much imagery as possible, depicting the four seasons within the typeface “” at the stroke-level. Reference 3 focuses on the shape distortion of typefaces at the multi-letter level. Some user design outcomes are shown in Fig. 7.

Task2: Creation (20 minutes). Participants are encouraged to explore and create their own designs freely. For those without a specific creative direction, we provide open-ended topics drawn from our collected corpus (cultural promotion, organizational branding, and personal identity) to inspire ideas. Thinking aloud during the creation process is encouraged, and some user design outcomes are shown in Fig. 8 and Fig. 9.

Interview (20 minutes). In the end, each participant completed a questionnaire with a 5-point Likert Scale. The questionnaire focused on three aspects of TypeDance: 1) the satisfaction of the generated outcome, 2) the usability of TypeDance system, and 3) the functionality of each individual component within the system. In addition, we conducted a semi-structured interview with each participant to collect their feedback on the design process.

Figure 8:

Figure 8: The second evaluation task is creation. For participants without a concrete creation goal, we provided open-ended topics as inspiration. Part of the open-ended topics design outcomes created by participants via TypeDance are shown, with annotations indicating the necessary design materials, including typeface and imagery, and the design patterns in their designs.

Figure 9:

Figure 9: In the creation task, some of the participants used TypeDance to fulfill their specific creative needs. For those participants, their personal information, scenario, and creation intent are recorded. Additionally, the necessary design materials, including typeface and imagery, along with the design outcomes are shown in the table.

7.3 Results Analysis

7.3.1 Satisfaction of Generated Outcome.

All Participants found that the generated outcome effectively blends both the information of the selected typeface and imagery (MEAN = 4.78, SD = 0.43), and the majority of them (MEAN = 4.17, SD = 0.62) agree the outcome can achieve a visually harmonious effect. Additionally, Over half of the participants (MEAN = 4.06, SD = 0.73) acknowledged that the generated outcomes were diverse. Their feedback supports that TypeDance is capable of achieving a natural blend and providing diverse results, which aligns with the second design consideration (D2) defined in Sect. 4.

Preservation. The majority of participants (11/18) expressed that the generated results were “beyond their expectation” and “innovative.” They found that TypeDance was capable of producing reasonable results that effectively combined both typeface and imagery. As mentioned by P3, “I initially didn’t see any relation between the swan and the letter ‘E,’ but the result showed that they could be combined in a way that is visually pleasing (P3, Fig. 7).”

Harmony. Participants (16/18) agreed that the generated results exhibited aesthetical harmony. TypeDance successfully maintained the legibility of the typeface while enhancing the visual appeal by incorporating imagery that “aligned with the skeleton of the text (P1, P4, Fig. 7).”

Diversity. Over half of the participants agreed that the generated results were diverse (14/18). Some participants (N=4) emphasized the importance of obtaining alternative designs in practice, commented that“Though I have achieved a satisfactory result, I still want to regenerate to see more interesting results (P2, Fig. 8; E7, Fig. 9).”

In terms of preservation , general users exhibited a lower sensitivity than designers in recognizing the typeface and imagery. In contrast, designers could swiftly perceive the content and showed a tendency to export potential designs to advanced tools for further preservation enhancement. Despite differing levels of design expertise, both novice users and designers demonstrated similar scores in terms of the harmony and diversity of the generated results. Besides perceptual harmony, designers identified more artistic effects. As E1 commented, “I never thought that AI could understand and produce negative space (Fig. 7).” In that case, TypeDance integrated the dog into the typeface by filling the empty space in the letter E” . During the creation process, all participants experimented with different combinations of design priors to achieve more diversified results. Interestingly, color was more frequently employed than shape, while semantics were consistently selected without specifying a text prompt.

Figure 10:

Figure 10: Users ratings for TypeDance, including (a) satisfaction of generated outcome, (b) usability of the system, and (c) usefulness of individual function. The above three stacked bar charts show the overall user rating for different indices of TypeDance, while the bellowing three range dot plots illustrate the distinct preferences between experts and novices, showcasing the mean rating values for these two user groups.

7.3.2 Usability of System.

The user study indicated that most participants (MEAN = 4.39, SD = 0.67) found TypeDance to maintain workflow integrity in the design process. Additionally, a majority (MEAN = 4.33, SD = 0.77) expressed satisfaction with the flexibility of blending different granularities of typeface and imagery. In terms of controllability during the generation process and editability of the generated result, more than half of the participants (N=12) agreed that TypeDance provided satisfactory control and editability options. These features align with the design considerations of customization and post-editability (D3 and D4) defined in Sect. 4.

Integrity. Most participants (N=10) strongly agreed the complete workflow has been instantiated within TypeDance. A participant highlighted, “I don’t need to switch between different platforms to finish a design (E2, Fig. 8).”

Flexibility. Half of the participants (N=9) strongly agreed with the flexibility provided by TypeDance to personalize their designs. Most participants (N=15) experimented with more than two types of typeface granularity in their designs. E2 pointed out, “I can easily select a single stroke that overlaps with other strokes in the typeface.”

Controllability. More than half of the participants (N=12) agreed that TypeDance provides a high level of controllability. They found that the generated results were able to accurately “reflect the selected imagery” and “adhere to the chosen shape”.

Editability. The post-editability of Typdance was strongly agreed upon by half of the participants (N=8). Several participants (N=3) expressed their desire for a generative tool that not only generates designs once but also provides the ability to make adjustments and rectify the results.

All participants widely recognized the workflow integrity of TypeDance, with different perspectives from designers and general users. Designers valued it for integrating essential functionalities that typically require switching between various platforms in the traditional workflow, while general users praised TypeDance for allowing them to sequentially follow the components in the interface to finish a design. Logo demands high customization with their special property of revealing identity. The option to select imagery from personal photos adds a personalized touch, surpassing the resources available in a shared community. Both designers and novices emphasized the ability to control and draw inspiration from the real world with specified visual representation, color, and shape. This feature is especially crucial in some scenarios, e.g. , “ designing a city logo.

The gap between flexibility and editability demonstrates the different expectations from designers and general users. General users demonstrated less interest in experimenting with different granularities of typeface, predominantly utilizing letter-level blending. Designers, on the other hand, highly praised this function as it allows them to segment various parts of the typeface or even combine across different granularities. After gaining the generated results, general users express satisfaction with changing colors or deleting elements (P4 & P6, Fig. 8). Designers find delight in the refinement function, as E5 notes “ it simulates the real design process where imagery is progressively simplified or details are added to the typeface (E5, Fig. 9). ” They also expressed a desire for more advanced editing functions, such as bezier curves, to fine-tune shapes."

7.3.3 Usefulness of Individual Functions.

Participants also provided evaluations for each component within the TypeDance system. The selection and generation components received unanimous agreement from all participants, with high and comparable scores. The coherence between selecting the typeface and imagery and considering design factors, such as preparing design materials, had a direct impact on the scores of the selection and generation components. E3 stated, “It saves much time for me to select the desired typeface and adjust the bezier curves to create shapes resembling specific objects, like a dog.” They also appreciated the diverse range of results offered by the system, which they found crucial for the design process (N=3). For the pre-generation, In terms of ideation, more than half of the participants (N=11) agreed that the concepts provided during the pre-generation phase are “ helpful to extend imagination” and “the explanation makes sense for me”. During the post-generation stage, the scores for evaluation and refinement were comparable due to the cohesive nature of the operations. Some participants (N=4, including E1, E3, and E4) expressed a particular satisfaction with these two components, as TypeDance achieves a “recognization the similarity between typeface and imagery” and “a more fine-grained adjustment that is independent of the generation”. These post-generation tools are “especially suitable for semantic typography design,” said by E1.

7.4 Limitation

In response to the issues encountered by users while using TypeDance, we identified the main limitations of the current system from three dimensions.

Figure 11:

Figure 11: The two tradeoffs revealed in participants’ cases. The figure on the left shows the tradeoff between imagery Diversity and style consistency, where the increasing number of imagery will lead to style variations. The figure on the right side shows the tradeoff between the typeface complexity and legibility, where the increased typeface complexity may lead to a potential decrease in the legibility of the results.

7.4.1 Trial and Error in Selecting Typeface and Imagery.

Although the current TypeDance allows creators to select and blend flexibly, facilitating quick generation, participant feedback suggests that trial-and-error with heterogeneous mapping between typeface and imagery could prolong the process. For instance, E9 achieved the final design after three attempts, experimenting with different typeface granularities, including “ Bea ”, single “ e ” and “a”, and “ ea ”. Participants noted that, apart from trying different parts of the typeface based on the selected imagery, it would also be challenging to find suitable imagery based on the chosen typeface."

7.4.2 Tradeoff between Imagery Diversity and Result Style Consistency.

Fig. 11 displays this limitation, where using the same imagery for “Hong Kong” produces stylistically consistent result, while using different imagery leads to noticeable inconsistency. E1, remarked, “These elements look good individually, but when combined, they appear discordant.” This inconsistency arises from transferring imagery and style from image references to the generated result, resulting in various styles when using multiple references. While adding a textual prompt is a partial solution to alleviate this issue, it lacks precise control. Incorporating multiple imageries within a single typeface is a common and significant format for semantic typographic logos. Thus, Achieving precise control over imagery diversity and result style consistency remains an important area for further investigation.

7.4.3 Tradeoff between Typeface Complexity and Result Legibility.

When the complexity of the typeface used in the creation rises (e.g., increasing the strokes or letters), the legibility of the generated result may decrease correspondingly. As the right side of Fig. 11 shows, using the same imagery during the creative process, the illustration of “” is easily readable, while the presentation of “” appears more abstract. It suggests that the current generation ability of TypeDance can not perform well in cases with complex typeface, e.g., an entire Chinese word, or multiple letters. The primary reason is that a complex typeface offers an initially rich structure for the generative model to stylize. To tackle this issue, TypeDance should find a viable solution to enhance control over the stylization progress during generation, achieving a balance between typeface complexity and legibility.

7.4.4 Image is not the perfect medium to eliminate the barrier of crafting a precise prompt.

To facilitate the creation of precise textual prompts, this study utilizes images with diverse design attributes in TypeDance. These serve as inspiration for users to select specific visual representations. However, images occasionally fail to meet users’ exact requirements due to the intricate interplay of layout, color, and semantic context within a single object. This complexity includes both desired and undesired elements, making it challenging for users to accurately convey their intentions. As highlighted by E7,“ While I observe three options for extracting distinct information from an image, opting for semantic extraction reveals that color still significantly influences the final outcome. ” P4 similarly remarks that finding an image perfectly aligned with her needs poses a challenge comparable to crafting a prompt from scratch.

Skip 8DISCUSSION Section

8 DISCUSSION

8.1 Personalized Design: Intent-aware Collaboration with AI

The rise of large language models has fueled a surge in text-driven creativity design [6, 51, 58], enabling creators to collaborate with AI using natural language narratives. While text-driven creation offers an intuitive means to manipulate the model in the backend without delving into complex parameters manually, expressing user intent concisely through textual prompts poses a challenge. Crafting a prompt becomes particularly daunting when describing an imagined visual design, given the myriad details such as layout, color, and shape that extend beyond textual representation. PromptPaint [10] recognizes this challenge and approaches it by mixing a set of textual prompts to capture ambiguous concepts, like “ a bit less vivid color. ” However, it remains constrained by offering a predefined set of prompts and fundamentally fails to resolve the issue of representing concrete visual concepts through prompts.

To ensure that a creator’s intention aligns seamlessly with AI collaboration, it is crucial to mirror real design practices with accessible design materials. The common design material used by creators includes explorable galleries [62], sketches [11], and even photographs capturing our perception of the world [31, 64]. These visual design materials encompass both explicit intentions, such as prominent semantics, and implicit aesthetic factors. In logo design, there is a pronounced emphasis on identity, using images frequently to convey intentions. This aspect can not be ignored in AI collaboration, demanding a capability for AI to comprehend visual semantics. It reveals that there is no universally superior material to encapsulate a creator’s intent; it depends on the design task. This necessitates a hybrid and multimodal collaboration that can flexibly generalize to a wide array of requirements.

8.2 Incorporating Design Knowledge into Creativity Support Tools

Instilling the generalizable design pattern into tools necessitates addressing technical and interaction challenges regarding how humans guide the model. AI models are often not built for the special design task, posing challenges in generalizing to complex patterns. For example, considerable research [47, 63] has delved into blending techniques for specific typeface granularity. However, creativity support tools are user-oriented, with more intricate design requirements, calling for advanced techniques to accommodate all levels of typeface granularity. Instead of retraining a model, significant research has explored to change or add the interaction with models for incorporating design knowledge, such as crowd-powered design parameterization [27] and intervention through intermediate representations [57].

The technical and interaction aspects of incorporating design knowledge externalize the idea of “ balancing automation and control, ” which is often noted by existing human-AI design guidelines [1, 2, 19]. The incorporation of design knowledge controlled by creators partially addresses the issue of AI copyright. Current generative models have faced criticism for sampling examples from the training set. In TypeDance, users contribute design materials through images, allowing for a personalized foundation instead of direct replication from a predefined dataset. This approach not only enhances creativity but also helps establish a stronger sense of ownership for creators. Complete automation with a single model to achieve an end-to-end result overlooks the user’s value. The allure of a creativity support tool, as opposed to relying solely on a model, lies in enabling creators to participate in crucial stages. This involvement includes customizing design materials, choosing which design knowledge to transfer to the generation process, and refining the final outcome.

8.3 Mix-User Oriented Design Workflow

With the aim of developing a tool with a “low threshold” for novices to steer the generation and a “high ceiling” for experts to achieve more advanced effects, TypeDance integrates a simulatable design workflow. Creativity support tools are inherently designed to provide a comprehensive authoring experience, addressing both common and unique emphases from diverse users [57, 58, 67]. As illustrated in Figure 10, experts and designers share both similar and distinct preferences. Functions with shared preferences, like selection and generation, can be considered central to the workflow and warrant deeper investigation. Notably, there are functions with varying preferences based on expertise. Experts tend to prioritize evaluation and refinement, whereas novices may view these as optional. However, with less design background, novices find ideation helpful than experts. Functions with differing preferences act as a “ wide wall ,” accommodating optional user requirements. While not mandatory like central functions, omitting them compromises the overall integrity of the workflow.

Skip 9CONCLUSION Section

9 CONCLUSION

This study distills design knowledge from real-world examples, summarizes generalizable design patterns and simulatable design workflow, and explores the creation of semantic typographic logos by blending typeface and imagery while maintaining legibility. We introduce TypeDance, an authoring tool based on a generative model that supports a personalized design workflow including ideation, selection, generation, evaluation, and iteration. With TypeDance, creators can flexibly choose typefaces at different levels of granularity and blend them with specific imagery using combinable design factors. TypeDance also allows users to adjust the generated results along the typeface-imagery spectrum and offers post-editing for individual elements. Feedback from general users and experts validates the effectiveness of TypeDance and provides valuable insights for future opportunities. We are excited to enhance the functionality of TypeDance for a comprehensive workflow and explore new techniques and interactions to enhance human creativity.

Skip ACKNOWLEDGMENTS Section

ACKNOWLEDGMENTS

The authors wish to thank the anonymous reviewers for their valuable comments. This work is supported in part by Guangzhou Basic and Applied Basic Research Foundation (2023A03J0142).

Footnotes

Skip Supplemental Material Section

Supplemental Material

Video Presentation

Video Presentation

mp4

94.6 MB

Video Introduction

Video Introduction

mp4

63.8 MB

References

  1. Saleema Amershi, Dan Weld, Mihaela Vorvoreanu, Adam Fourney, Besmira Nushi, Penny Collisson, Jina Suh, Shamsi Iqbal, Paul N Bennett, Kori Inkpen, 2019. Guidelines for human-AI interaction. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1–13. https://doi.org/10.1145/3290605.3300233Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Apple. 2021. Human Interface Guidelines. https://developer.apple.com/design/human-interface-guidelines/, Last accessed on 2023-12-13.Google ScholarGoogle Scholar
  3. Gregory Ashworth and Mihalis Kavaratzis. 2009. Beyond the logo: Brand management for cities. Journal of Brand Management 16 (2009), 520–531. https://doi.org/10.1057/palgrave.bm.2550133Google ScholarGoogle ScholarCross RefCross Ref
  4. Daniel Berio, Frederic Fol Leymarie, Paul Asente, and Jose Echevarria. 2022. Strokestyles: Stroke-based segmentation and stylization of fonts. ACM Transactions on Graphics 41, 3 (2022), 21 pages. https://doi.org/10.1145/3505246Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. James Betker, Gabriel Goh, Li Jing, Tim Brooks, Jianfeng Wang, Linjie Li, Long Ouyang, Juntang Zhuang, Joyce Lee, Yufei Guo, 2023. Improving image generation with better captions. Computer Science. https://cdn. openai. com/papers/dall-e-3. pdf (2023).Google ScholarGoogle Scholar
  6. Yining Cao, Jane L E, Zhutian Chen, and Haijun Xia. 2023. DataParticles: Block-based and language-oriented authoring of animated unit visualizations. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1–15. https://doi.org/10.1145/3544548.3581472Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Terry L Childers and Jeffrey Jass. 2002. All dressed up with something to say: Effects of typeface semantic associations on brand perceptions and consumer memory. Journal of Consumer Psychology 12, 2 (2002), 93–106. https://doi.org/10.1207/S15327663JCP1202_03Google ScholarGoogle ScholarCross RefCross Ref
  8. Lydia B Chilton, Ecenaz Jen Ozmen, Sam H Ross, and Vivian Liu. 2021. VisiFit: Structuring iterative improvement for novice designers. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1–14. https://doi.org/10.1145/3411764.3445089Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Lydia B Chilton, Savvas Petridis, and Maneesh Agrawala. 2019. VisiBlends: A flexible workflow for visual blends. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1–14. https://doi.org/10.1145/3290605.3300402Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. John Joon Young Chung and Eytan Adar. 2023. PromptPaint: Steering Text-to-Image Generation Through Paint Medium-like Interactions. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology. ACM, 1–17. https://doi.org/10.1145/3586183.3606777Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. John Joon Young Chung, Wooseok Kim, Kang Min Yoo, Hwaran Lee, Eytan Adar, and Minsuk Chang. 2022. TaleBrush: Sketching stories with generative pretrained language models. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1–19. https://doi.org/10.1145/3491102.3501819Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Weiwei Cui, Xiaoyu Zhang, Yun Wang, He Huang, Bei Chen, Lei Fang, Haidong Zhang, Jian-Guan Lou, and Dongmei Zhang. 2019. Text-to-viz: Automatic generation of infographics from proportion-related natural language statements. IEEE Transactions on Visualization and Computer Graphics 26, 1 (2019), 906–916. https://doi.org/10.1109/TVCG.2019.2934785Google ScholarGoogle ScholarCross RefCross Ref
  13. João M Cunha, Nuno Lourenço, Pedro Martins, and Penousal Machado. 2020. Visual blending for concept representation: A case study on emoji generation. New Generation Computing 38, 4 (2020), 739–771. https://doi.org/10.1007/s00354-020-00107-xGoogle ScholarGoogle ScholarDigital LibraryDigital Library
  14. Laura Devendorf and Kimiko Ryokai. 2013. AnyType: provoking reflection and exploration with aesthetic interaction. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1041–1050. https://doi.org/10.1145/2470654.2466133Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Ryan Dew, Asim Ansari, and Olivier Toubia. 2022. Letting logos speak: Leveraging multiview representation learning for data-driven branding and logo design. Marketing Science 41, 2 (2022), 401–425. https://doi.org/10.1287/mksc.2021.1326Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Evelyn Fix and Joseph Lawson Hodges. 1989. Discriminatory analysis. Nonparametric discrimination: Consistency properties. International Statistical Review/Revue Internationale de Statistique 57, 3 (1989), 238–247.Google ScholarGoogle ScholarCross RefCross Ref
  17. Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H. Bermano, Gal Chechik, and Daniel Cohen-Or. 2022. An Image is Worth One Word: Personalizing Text-to-Image Generation using Textual Inversion. arXiv preprint arXiv:2208.01618.Google ScholarGoogle Scholar
  18. Rinon Gal, Or Patashnik, Haggai Maron, Amit H Bermano, Gal Chechik, and Daniel Cohen-Or. 2022. StyleGAN-NADA: CLIP-guided domain adaptation of image generators. ACM Transactions on Graphics 41, 4, Article 141 (2022), 13 pages. https://doi.org/10.1145/3528223.3530164Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Google. 2019. People + AI Guidebook. https://pair.withgoogle.com/, Last accessed on 2023-12-13.Google ScholarGoogle Scholar
  20. Pamela W Henderson and Joseph A Cote. 1998. Guidelines for selecting or modifying logos. Journal of Marketing 62, 2 (1998), 14–30. https://doi.org/10.2307/1252158Google ScholarGoogle ScholarCross RefCross Ref
  21. Yon Ade Lose Hermanto. 2023. Semantic Interpretation in Experimental Typography Creation. KnE Social Sciences 8, 15 (2023), 252–257. https://doi.org/10.18502/kss.v8i15.13939Google ScholarGoogle ScholarCross RefCross Ref
  22. Kai-Wen Hsiao, Yong-Liang Yang, Yung-Chih Chiu, Min-Chun Hu, Chih-Yuan Yao, and Hung-Kuo Chu. 2023. Img2Logo: Generating Golden Ratio Logos from Images. In Computer Graphics Forum, Vol. 42. Wiley Online Library, 37–49. https://doi.org/10.1111/cgf.14742Google ScholarGoogle ScholarCross RefCross Ref
  23. Shir Iluz, Yael Vinker, Amir Hertz, Daniel Berio, Daniel Cohen-Or, and Ariel Shamir. 2023. Word-As-Image for Semantic Typography. ACM Transactions on Graphics 42, 4, Article 151 (2023), 11 pages. https://doi.org/10.1145/3592123Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Youwen Kang, Zhida Sun, Sitong Wang, Zeyu Huang, Ziming Wu, and Xiaojuan Ma. 2021. MetaMap: Supporting visual metaphor ideation through multi-dimensional example-based exploration. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 15 pages. https://doi.org/10.1145/3411764.3445325Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Alexander Kirillov, Eric Mintun, Nikhila Ravi, Hanzi Mao, Chloe Rolland, Laura Gustafson, Tete Xiao, Spencer Whitehead, Alexander C. Berg, Wan-Yen Lo, Piotr Dollár, and Ross Girshick. 2023. Segment Anything. arXiv preprint arXiv:2304.02643 (2023).Google ScholarGoogle Scholar
  26. Janin Koch, Andrés Lucero, Lena Hegemann, and Antti Oulasvirta. 2019. May AI? Design ideation with cooperative contextual bandits. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1–12. https://doi.org/10.1145/3290605.3300863Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Yuki Koyama, Daisuke Sakamoto, and Takeo Igarashi. 2014. Crowd-powered parameter analysis for visual design exploration. In Proceedings of the 27th annual ACM symposium on User interface software and technology. ACM, 65–74. https://doi.org/10.1145/2642918.2647386Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Jieun Lee, Eunju Ko, and Carol M Megehee. 2015. Social benefits of brand logos in presentation of self in cross and same gender influence contexts. Journal of Business Research 68, 6 (2015), 1341–1349. https://doi.org/10.1016/j.jbusres.2014.12.004Google ScholarGoogle ScholarCross RefCross Ref
  29. Junnan Li, Dongxu Li, Caiming Xiong, and Steven Hoi. 2022. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In Proceedings of the International Conference on Machine Learning, Vol. 162. PMLR, 12888–12900.Google ScholarGoogle Scholar
  30. Yi-Na Li, Kang Zhang, and Dong-Jin Li. 2017. Rule-based automatic generation of logo designs. Leonardo 50, 2 (2017), 177–181. https://doi.org/10.1162/LEON_a_00961Google ScholarGoogle ScholarCross RefCross Ref
  31. Giorgia Lupi and Stefanie Posavec. 2018. Observe, collect, draw!: A visual journal: Discover the patterns in your everyday life. Princeton Architectural Press.Google ScholarGoogle Scholar
  32. O Mataev and H Mataev. 2006. Olga’s gallery. giuseppe Arcimboldo.Google ScholarGoogle Scholar
  33. Mark Meyer, Alan Barr, Haeyoung Lee, and Mathieu Desbrun. 2002. Generalized barycentric coordinates on irregular polygons. Journal of Graphics Tools1 (2002), 13–22.Google ScholarGoogle Scholar
  34. Chong Mou, Xintao Wang, Liangbin Xie, Yanze Wu, Jian Zhang, Zhongang Qi, Ying Shan, and Xiaohu Qie. 2023. T2i-adapter: Learning adapters to dig out more controllable ability for text-to-image diffusion models. arXiv preprint arXiv:2302.08453 (2023).Google ScholarGoogle Scholar
  35. Takeshi Okada and Kentaro Ishibashi. 2017. Imitation, inspiration, and creation: Cognitive process of creative drawing by copying others’ artworks. Cognitive Science 41, 7 (2017), 1804–1837. https://doi.org/10.1111/cogs.12442Google ScholarGoogle ScholarCross RefCross Ref
  36. Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems 35 (2022), 27730–27744.Google ScholarGoogle Scholar
  37. Helen Petrie, Fraser Hamilton, and Neil King. 2004. Tension, what tension? Website accessibility and visual design. In Proceedings of the International Cross-Disciplinary Workshop on Web Accessibility, Vol. 63. ACM, 13–18. https://doi.org/10.1145/990657.990660Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. Huy Quoc Phan, Hongbo Fu, and Antoni B Chan. 2015. Flexyfont: Learning transferring rules for flexible typeface synthesis. In Computer Graphics Forum, Vol. 34. Wiley Online Library, 245–256. https://doi.org/10.1111/cgf.12763Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, 2021. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Vol. 139. PMLR, 8748–8763.Google ScholarGoogle Scholar
  40. Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. 2022. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125 (2022).Google ScholarGoogle Scholar
  41. Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, and Ilya Sutskever. 2021. Zero-shot text-to-image generation. In Proceedings of the International Conference on Machine Learning, Vol. 139. PMLR, 8821–8831.Google ScholarGoogle Scholar
  42. Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. 2022. High-Resolution Image Synthesis With Latent Diffusion Models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. IEEE, 10684–10695. https://doi.org/10.1109/CVPR52688.2022.01042Google ScholarGoogle ScholarCross RefCross Ref
  43. Subhadip Roy and Rekha Attri. 2022. Physimorphic vs. Typographic logos in destination marketing: Integrating destination familiarity and consumer characteristics. Tourism Management 92 (2022), 104544. https://doi.org/10.1016/j.tourman.2022.104544Google ScholarGoogle ScholarCross RefCross Ref
  44. Patsorn Sangkloy, Wittawat Jitkrittum, Diyi Yang, and James Hays. 2022. A sketch is worth a thousand words: Image retrieval with text and sketch. In Proceedings of the European Conference on Computer Vision. Springer, 251–267. https://doi.org/10.1007/978-3-031-19839-7_15Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. Yang Shi, Pei Liu, Siji Chen, Mengdi Sun, and Nan Cao. 2022. Supporting expressive and faithful pictorial visualization design with visual style transfer. IEEE Transactions on Visualization and Computer Graphics 29, 1 (2022), 236–246. https://doi.org/10.1109/TVCG.2022.3209486Google ScholarGoogle ScholarCross RefCross Ref
  46. Ben Shneiderman. 2007. Creativity support tools: accelerating discovery and innovation. Commun. ACM 50, 12 (2007), 20–32. https://doi.org/10.1145/1323688.1323689Google ScholarGoogle ScholarDigital LibraryDigital Library
  47. Maham Tanveer, Yizhi Wang, Ali Mahdavi-Amiri, and Hao Zhang. 2023. DS-Fusion: Artistic Typography via Discriminated and Stylized Diffusion. In Proceedings of the International Conference on Computer Vision. IEEE, 374–384. https://doi.org/10.1109/ICCV51070.2023.00041Google ScholarGoogle ScholarCross RefCross Ref
  48. Purva Tendulkar, Kalpesh Krishna, Ramprasaath R Selvaraju, and Devi Parikh. 2019. Trick or TReAT: Thematic reinforcement for artistic typography. In International Conference on Computer and Communications. arXiv:1903.07820Google ScholarGoogle Scholar
  49. Dimitri von Rütte, Elisabetta Fedele, Jonathan Thomm, and Lukas Wolf. 2023. FABRIC: Personalizing Diffusion Models with Iterative Feedback. arXiv preprint arXiv:2307.10159 (2023).Google ScholarGoogle Scholar
  50. Yizhi Wang, Yue Gao, and Zhouhui Lian. 2020. Attribute2font: Creating fonts you want from attributes. ACM Transactions on Graphics 39, 4, Article 69 (2020), 15 pages. https://doi.org/10.1145/3386569.3392456Google ScholarGoogle ScholarDigital LibraryDigital Library
  51. Yun Wang, Zhitao Hou, Leixian Shen, Tongshuang Wu, Jiaqi Wang, He Huang, Haidong Zhang, and Dongmei Zhang. 2022. Towards natural language-based visualization authoring. IEEE Transactions on Visualization and Computer Graphics 29, 1 (2022), 1222–1232. https://doi.org/10.1109/TVCG.2022.3209357Google ScholarGoogle ScholarCross RefCross Ref
  52. Yizhi Wang, Guo Pu, Wenhan Luo, Yexin Wang, Pengfei Xiong, Hongwen Kang, and Zhouhui Lian. 2022. Aesthetic text logo synthesis via content-aware layout inferring. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. IEEE, 2436–2445. https://doi.org/arXiv:2204.02701v1Google ScholarGoogle ScholarCross RefCross Ref
  53. Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V Le, Denny Zhou, 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems 35 (2022), 24824–24837.Google ScholarGoogle Scholar
  54. Shishi Xiao, Suizi Huang, Yue Lin, Yilin Ye, and Wei Zeng. 2024. IEEE Transactions on Visualization and Computer Graphics 30, 1 (2024), 284–294. https://doi.org/10.1109/TVCG.2023.3326913Google ScholarGoogle ScholarDigital LibraryDigital Library
  55. Jie Xu and Craig S Kaplan. 2007. Calligraphic packing. In Proceedings of Graphics Interface. ACM, 43–50. https://doi.org/10.1145/1268517.1268527Google ScholarGoogle ScholarDigital LibraryDigital Library
  56. Xiaotong Xu, Rosaleen Xiong, Boyang Wang, David Min, and Steven P Dow. 2021. Ideaterelate: An examples gallery that helps creators explore ideas in relation to their own. Proceedings of the ACM on Human-Computer Interaction 5, CSCW2, Article 352 (2021), 18 pages. https://doi.org/10.1145/3479496Google ScholarGoogle ScholarDigital LibraryDigital Library
  57. Chuan Yan, John Joon Young Chung, Yoon Kiheon, Yotam Gingold, Eytan Adar, and Sungsoo Ray Hong. 2022. FlatMagic: Improving flat colorization through AI-driven design for digital comic professionals. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1–17. https://doi.org/10.1145/3491102.3502075Google ScholarGoogle ScholarDigital LibraryDigital Library
  58. Zihan Yan, Chunxu Yang, Qihao Liang, and Xiang’Anthony’ Chen. 2023. XCreation: A Graph-based Crossmodal Generative Creativity Support Tool. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology. ACM, 1–15. https://doi.org/10.1145/3586183.3606826Google ScholarGoogle ScholarDigital LibraryDigital Library
  59. Shuai Yang, Jiaying Liu, Zhouhui Lian, and Zongming Guo. 2017. Awesome typography: Statistics-based text effects transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. IEEE, 7464–7473. https://doi.org/10.1109/CVPR.2017.308Google ScholarGoogle ScholarCross RefCross Ref
  60. Lin-Ping Yuan, Wei Zeng, Siwei Fu, Zhiliang Zeng, Haotian Li, Chi-Wing Fu, and Huamin Qu. 2021. Deep colormap extraction from visualizations. IEEE Transactions on Visualization and Computer Graphics 28, 12 (2021), 4048–4060. https://doi.org/10.1109/TVCG.2021.3070876Google ScholarGoogle ScholarDigital LibraryDigital Library
  61. Lin-Ping Yuan, Ziqi Zhou, Jian Zhao, Yiqiu Guo, Fan Du, and Huamin Qu. 2021. Infocolorizer: Interactive recommendation of color palettes for infographics. IEEE Transactions on Visualization and Computer Graphics 28, 12 (2021), 4252–4266. https://doi.org/10.1109/TVCG.2021.3085327Google ScholarGoogle ScholarDigital LibraryDigital Library
  62. Enhao Zhang and Nikola Banovic. 2021. Method for Exploring Generative Adversarial Networks (GANs) via Automatically Generated Image Galleries. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1–15. https://doi.org/10.1145/3411764.3445714Google ScholarGoogle ScholarDigital LibraryDigital Library
  63. Junsong Zhang, Yu Wang, Weiyi Xiao, and Zhenshan Luo. 2017. Synthesizing ornamental typefaces. In Computer Graphics Forum, Vol. 36. Wiley Online Library, 64–75. https://doi.org/10.1111/cgf.12785Google ScholarGoogle ScholarDigital LibraryDigital Library
  64. Jiayi Eris Zhang, Nicole Sultanum, Anastasia Bezerianos, and Fanny Chevalier. 2020. DataQuilt: Extracting visual elements from images to craft pictorial visualizations. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1–13. https://doi.org/10.1145/3313831.3376172Google ScholarGoogle ScholarDigital LibraryDigital Library
  65. Lvmin Zhang, Anyi Rao, and Maneesh Agrawala. 2023. Adding conditional control to text-to-image diffusion models. In Proceedings of the International Conference on Computer Vision. 3836–3847. https://doi.org/10.1109/ICCV51070.2023.00355Google ScholarGoogle ScholarCross RefCross Ref
  66. Nanxuan Zhao, Nam Wook Kim, Laura Mariah Herman, Hanspeter Pfister, Rynson WH Lau, Jose Echevarria, and Zoya Bylinskii. 2020. Iconate: Automatic compound icon generation and ideation. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1–13. https://doi.org/10.1145/3313831.3376618Google ScholarGoogle ScholarDigital LibraryDigital Library
  67. Tongyu Zhou, Connie Liu, Joshua Kong Yang, and Jeff Huang. 2023. filtered. ink: Creating Dynamic Illustrations with SVG Filters. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 1–15. https://doi.org/10.1145/3544548.3581051Google ScholarGoogle ScholarDigital LibraryDigital Library
  68. Changqing Zou, Junjie Cao, Warunika Ranaweera, Ibraheem Alhashim, Ping Tan, Alla Sheffer, and Hao Zhang. 2016. Legible compact calligrams. ACM Transactions on Graphics 35, 4, Article 122 (2016), 12 pages. https://doi.org/10.1145/2897824.2925887Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. TypeDance: Creating Semantic Typographic Logos from Image through Personalized Generation

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          CHI '24: Proceedings of the CHI Conference on Human Factors in Computing Systems
          May 2024
          18961 pages
          ISBN:9798400703300
          DOI:10.1145/3613904

          Copyright © 2024 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 11 May 2024

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article
          • Research
          • Refereed limited

          Acceptance Rates

          Overall Acceptance Rate6,199of26,314submissions,24%
        • Article Metrics

          • Downloads (Last 12 months)207
          • Downloads (Last 6 weeks)207

          Other Metrics

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        HTML Format

        View this article in HTML Format .

        View HTML Format