1. Introduction
Motion generation, a pivotal aspect of computer vision and animation, aims to create life-like human movements for applications ranging from virtual reality to interactive gaming. This field has evolved from leveraging basic kinematic models to adopting deep learning techniques, significantly enhancing the realism and dynamism of generated motions. The incorporation of large motion capture datasets and neural networks has enabled the synthesis of complex movements from a variety of inputs, such as music and textual descriptions. Despite these advances, generating semantically rich and physically plausible motions remains a challenging endeavor, necessitating innovative approaches to bridge the gap between motion dynamics and input conditions [
1,
2,
3,
4].
Recent studies in motion generation have showcased impressive capabilities in creating dance movements from music and actions from textual descriptions. However, these methods often rely on extensive motion capture datasets, which are costly and labor-intensive to produce. Furthermore, the generalization of these models to open-vocabulary texts and the inclusion of emotional nuances in generated motions pose significant challenges. The reliance on specific annotations limits the diversity of achievable motions, and the lack of emotional expressiveness in generated movements restricts the applicability of these technologies in creating truly immersive experiences [
5,
6,
7].
Inspired by the success of pretrained models in vision and language tasks, such as CLIP, we explore the potential of leveraging these advancements for motion generation. The ability of models like CLIP to align the semantic spaces of language and vision hints at the possibility of similarly aligning textual descriptions with motion dynamics. This approach promises to address the challenges of open-vocabulary motion generation and the need for diverse and emotionally expressive movements, paving the way for generating more nuanced and context-aware human motions [
8,
9,
10].
We propose MLUG, a novel framework for motion generation that draws inspiration from BLIP’s methodologies. MLUG consists of a unimodal encoder trained with a motion–text contrastive loss, a motion-grounded text encoder for modeling motion–language interactions, and a motion-grounded motion decoder for generating motions from textual descriptions. Additionally, MLUG incorporates a motion length predictor to estimate the appropriate duration of generated motions based on the given text, addressing the limitations of previous models in generating semantically aligned and emotionally expressive motions [
2,
3].
Our evaluations demonstrate MLUG’s effectiveness in generating realistic and diverse motions from a wide range of textual descriptions, outperforming existing motion generation models in terms of realism, diversity, and semantic alignment. By leveraging the principles of vision–language pre-training, MLUG marks a significant step forward in synthesizing human motions that are both physically plausible and emotionally resonant.
In summary, our contributions are as follows:
Introduction of MLUG, a novel motion generation framework that integrates the advancements of vision–language pre-training to overcome the challenges of open-vocabulary motion generation and emotional expressiveness.
A comprehensive architecture that includes a unimodal encoder, a motion-grounded text encoder, a motion-grounded motion decoder, and a motion length predictor, enabling the generation of nuanced and contextually relevant motions.
Extensive evaluations showcasing MLUG’s superior performance in generating high-quality motions, significantly advancing the state of the art in motion generation.
2. MLUG
As shown in
Figure 1, MLUG consists of four training parts: MTC, MTM, MGM, and MLM. The MLUG (Motion Language Understanding and Generation) framework aims to bridge the gap between textual descriptions and human motion generation by integrating advanced NLP techniques with motion synthesis. By leveraging a multimodal approach, MLUG not only generates motions that are semantically aligned with textual inputs but also ensures the physical plausibility and emotional expressiveness of the generated movements.
2.1. Unimodal Encoder with Motion–Text Contrastive (MTC) Loss
In the development of MLUG, our first endeavor was to tackle the challenge of semantic alignment between text and motion. The unimodal encoder emerged as a cornerstone, designed to distill textual descriptions into a rich feature space. Drawing inspiration from the success of contrastive learning in vision–language pre-training, we introduced the motion–text contrastive (MTC) loss. This choice was motivated by the potential of contrastive learning to enhance the encoder’s ability to discern and align with the nuanced semantics of motion described in text.
where
and
denote the encoded features of motion and text, respectively,
is a similarity function,
is a temperature parameter, and
N is the number of negative samples.
2.2. Motion-Grounded Text Encoder with Motion–Text Matching (MTM) Loss
The journey continued with enhancing MLUG’s comprehension of the complex interplay between motion and language. The motion-grounded text encoder was a pivotal innovation, designed to delve deeper into this relationship. The introduction of the motion–text matching (MTM) loss marked a significant step forward, enabling the encoder to not just encode, but truly understand and distinguish between congruent and incongruent motion–text pairs.
where
s is the score assigned by the encoder to a motion–text pair,
denotes the sigmoid function, and
y is the ground truth label indicating whether the pair matches.
2.3. Motion-Grounded Motion Decoder with Motion Modeling (MGM) Loss
The motion-grounded motion decoder represents the culmination of MLUG’s capabilities, turning understanding into creation. Adopting a causal self-attention mechanism, this component was meticulously crafted to ensure that the generated motions not only align with the textual descriptions but also flow in a temporally coherent manner. The motion modeling (MGM) loss was introduced to refine the decoder’s output, focusing on the fidelity and fluidity of the generated motions.
where
and
represent the ground truth and predicted motion vectors at time step
i, respectively, and
T is the total number of time steps.
2.4. Motion Length Predictor (MLP)
As shown in
Figure 2, motions with similar descriptions are normally similar. Thus, we use retrieved motion to assist predict the length of generated motion here. The motion length predictor (MLP) emerged from the realization that the complexity of textual descriptions often implies a corresponding variability in motion length. The MLP was designed to tackle this novel challenge, employing a softmax layer to predict the appropriate duration of generated motions based on textual input.
where
L is the predicted length,
is the text encoding,
and
are the weights and bias of the MLP, respectively.
2.5. Integration of Components
The integration of these components into a cohesive system was a critical phase in MLUG’s development. The overall objective function reflects a balanced synthesis of the insights and innovations from each component, capturing the essence of MLUG’s multidimensional approach to motion generation.
where
,
,
, and
are hyperparameters to weight the importance of each loss component.
3. Training and Optimization of MLUG
The training and optimization process of MLUG is pivotal in enabling it to generate human motions that are semantically aligned with textual inputs and exhibit temporal coherence. This section discusses the comprehensive strategy employed, highlighting the synergy between different loss functions, optimization strategies, and the deployment of advanced techniques to fine-tune the model’s efficacy.
3.1. Adaptive Learning Rate Scheduling
To navigate the intricate architecture of MLUG and its diverse training goals, we adopt an adaptive learning rate scheduler that modifies the learning rate based on validation loss performance dynamically:
Here, signifies the learning rate at training step t, is the initial learning rate, T denotes the total number of training steps, and indicates the step interval at which the learning rate decay factor is halved, promoting a cyclical yet exponentially diminishing pattern in the learning rate adjustment.
3.2. Joint Optimization of Loss Functions
A hallmark of MLUG’s training regimen is the concurrent optimization of several loss functions, each tailored to a specific aspect of motion generation. The cumulative loss,
, amalgamates the motion–text contrastive (MTC) loss, motion–text matching (MTM) loss, motion modeling (MGM) loss, and motion length prediction (MLP) loss:
In this equation, , , , and serve as hyperparameters that modulate the influence of each loss component, with the optimization goal being the minimization of through gradient descent and backpropagation.
3.3. Stochastic Gradient Descent with Momentum
To adeptly explore MLUG’s complex loss surface, Stochastic Gradient Descent (SGD) with momentum is utilized. This method not only expedites convergence but also aids in averting entrapment in local minima:
Here, denotes the momentum term at training step t, symbolizes the momentum coefficient, represents the model parameters at step t, and is the gradient of the total loss with respect to the parameters.
3.4. Regularization Techniques
Given MLUG’s complexity and the motion data’s high-dimensional nature, regularization techniques are crucial for preventing overfitting and enhancing the model’s generalization capabilities. We integrate dropout and weight decay as part of our regularization strategy, fine-tuning their parameters to balance between model complexity and fidelity to the training data.
3.5. Batch Normalization for Stabilization
Batch normalization layers are incorporated within MLUG’s architecture to further stabilize the training process and facilitate faster convergence. This approach normalizes the inputs to each layer across the batch to have zero mean and unit variance, thus reducing internal covariate shift:
where
is the input to a batch normalization layer for the
k-th feature,
and
are the batch’s mean and variance, and
is a small constant for numerical stability.
The outlined training and optimization strategy for MLUG meticulously harnesses the architecture’s potential, ensuring the generation of semantically aligned and temporally coherent human motions.
4. Experiment
Extensive comparisons evaluate the performance of our MLUG across multiple motion-relevant tasks and datasets. In our experiments, we introduce details of the dataset settings, evaluation metrics, and implementation specifics, compared results with SOTAs, ablation study and case study (
Table 1).
4.1. Experimental Setup
Datasets. General motion synthesis can support diverse task settings, and thus previous datasets and a modified benchmark are utilized to evaluate MLUG. The study primarily focuses on two text-to-motion datasets: HumanML3D [
11] and KIT [
12]. The KIT dataset provides 6353 textual descriptions corresponding to 3911 motion sequences, while the HumanML3D dataset [
11] is a more recent dataset that contains 14,616 motion sequences obtained from AMASS [
13], along with 44,970 sequence-level textual descriptions. To evaluate MLUG as a uniform framework on tasks, such as motion prediction, we utilize the motion sequences available in HumanML3D, which is also a subset of the larger AMASS dataset. Following the previous works [
11,
14,
15], we adopt the same motion representation for fair comparisons, which combines joint velocities, positions, and rotations. By using this consistent representation, MLUG enables the availability to support further studies in the field.
Evaluation Metrics: (1) Motion quality: Frechet Inception Distance (FID) is our primary metric based on a feature extractor [
11] to evaluate the distance of feature distributions between the generated and real motions. For motion completion, we utilize metrics used in motion prediction studies [
16,
17,
18], such as Average Displacement Error (ADE) and Final Displacement Error (FDE), to evaluate the accuracy of the predicted motion.
(2) Generation diversity: We utilize the Diversity (DIV) metric to assess the motions diversity, which calculates the variance through features extracted from the motions [
11]. MultiModality (MM) measures the diversity of generated motions within the same text description of motion. (3) Text matching: Based on the feature space from [
11], the motion-retrieval precision (R Precision) evaluates the accuracy of matching between texts and motions using Top 1/2/3 retrieval accuracy. Multi-modal Distance (MM Dist) measures the distance between motions and texts. (4) Linguistic quality: We follow [
19] utilizing linguistic metrics from natural language studies, including BLUE [
20], Rouge [
21], Cider [
22], and BertScore [
23] to evaluate the quality of generated motion captions.
Implementation Details: We set the codebook of motion tokenizer as for most comparisons. The motion encoder incorporates a temporal downsampling rate l of 4. s The feed-forward networks have an output dimensionality of , and the attention mechanisms employ an inner dimensionality of . The remaining sub-layers and embeddings have a dimensionality of .
Table 1.
Comparison of four motion-related tasks on HumanML3D [
11] dataset. The evaluation metrics are computed using the encoder introduced in [
24]. The empty columns of previous methods indicate that they can not handle the task. The arrows (→) indicate that closer to
Real is desirable.
Bold and
underline indicate the best and the second best result on text-to-motion task.
Table 1.
Comparison of four motion-related tasks on HumanML3D [
11] dataset. The evaluation metrics are computed using the encoder introduced in [
24]. The empty columns of previous methods indicate that they can not handle the task. The arrows (→) indicate that closer to
Real is desirable.
Bold and
underline indicate the best and the second best result on text-to-motion task.
Methods | Text-to-Motion | Motion-to-Text | Motion Prediction | Motion In-Between |
---|
R TOP1↑
|
FID↓
|
DIV→
|
R TOP3↑
|
Bleu@4↑
|
Cider↑
|
FID↓
|
DIV→
|
FID↓
|
DIV→
|
---|
Real | | | | | - | - | 0.002 | 9.503 | 0.002 | 9.503 |
MLD [14] | | | | - | - | - | - | - | - | - |
T2M-GPT [15] | | | | - | - | - | - | - | - | - |
TM2T [19] | | | | | | | - | - | - | - |
MDM [15] | | | | - | - | - | | | | |
MLUG (Ours) | | | | | | | | | | |
4.2. Comparisons on Text-to-Motion on Human3D
In the evaluation of text-to-motion generation, MLUG emerges as the standout method, demonstrating superior performance across all metrics when compared against a selection of prior approaches on the HumanML3D dataset (
Table 2). Specifically, MLUG achieves a remarkable R TOP1 score of
, clearly outperforming its closest competitor, T2M-GPT, which scores
. This achievement underscores MLUG’s effectiveness in generating highly relevant motions corresponding to textual descriptions. Moreover, in terms of FID, MLUG significantly outshines other methods with a score of
, with T2M-GPT trailing behind with a score of
. This lower FID score for MLUG indicates its capability to produce motions that are not only diverse but also closely resemble the real data distribution, a testament to the method’s quality and diversity in generated motions. Additionally, MLUG sets a new benchmark in diversity (DIV) with a score of
, closely mirroring the real data’s diversity score of
. This comparison highlights MLUG’s unparalleled ability to generate a wide variety of motions that closely match the natural motion diversity found in the real-world data. The significant performance gap between MLUG and other methods, such as MLD and TM2T, further cements MLUG’s status as the leading approach for text-to-motion generation. MLD, for instance, presents an R TOP1 score of
and an FID of
, both of which fall short when compared to MLUG’s scores. Similarly, TM2T shows a lower performance with an R TOP1 score of
and an FID of
, indicating a wider gap from the real data distribution compared to MLUG.
4.3. Comparisons on Text-to-Motion on KIT
In evaluating text-to-motion generation on the KIT dataset, the performance of various methods is closely examined across multiple metrics: RPrecision (Top1, Top2, Top3), FID, MMDist, diversity, and MModality. The “Real” data sets a benchmark with RPrecision scores at for Top1, for Top2, and for Top3, alongside an FID of , MMDist of , and a diversity score of .
Among the compared methods, T2M-GPT comes closest to the real data’s RPrecision scores, with Top1 at , Top2 at , and Top3 at . This indicates a high relevance of generated motions to the text descriptions. However, its FID score of and MMDist of suggest a slight compromise in motion quality and distribution match to the real dataset.
MLD shows commendable performance with RPrecision scores nearly matching those of real data and significantly outperforming other methods with an FID of and MMDist of . This suggests that MLD effectively balances motion relevance to text with high-quality motion generation.
4.4. Ablation Study
To evaluate the effectiveness of different training strategies, we design our variants of MLUG as , , , and , where means “without”.
Ablation Study on HumanML3D
As shown in
Table 3, the ablation study on the HumanML3D dataset provides insightful observations on the impact of various components within the MLUG framework. Removing specific modules from MLUG—namely motion–text consistency (MTC), motion–text matching (MTM), motion generation module (MGM), and motion–language modeling (MLM)—yields variations in performance across different tasks, including text-to-motion, motion-to-text, motion prediction, and motion in-between. Notably, the omission of the MGM component leads to a slight decrease in diversity (DIV) and an increase in Frechet Inception Distance (FID), underscoring its critical role in enhancing the variety and quality of generated motions. Conversely, the removal of the MLM component exhibits a significant impact on the linguistic alignment, evidenced by the drop in Bleu@4 and Cider scores, highlighting the importance of language understanding in the generation process. These findings underscore the integral contributions of each module to the holistic performance of MLUG, emphasizing that the synergistic interaction between motion and text components is paramount for achieving state-of-the-art results in human motion generation tasks.
Ablation Study on KIT
As shown in
Table 4, the ablation study conducted on the KIT dataset reveals the nuanced impact of various components on the MLUG model’s performance in text-driven motion generation tasks. By systematically removing key modules—motion–text consistency (MTC), motion–text matching (MTM), motion generation module (MGM), and motion–language modeling (MLM)—we observe distinct shifts in model efficacy across a spectrum of metrics. Notably, the removal of the MGM module slightly degrades the model’s FID and diversity scores, indicating its pivotal role in generating varied and high-fidelity motions. Conversely, omitting the MLM component affects the RPrecision scores, highlighting the importance of language understanding in accurately generating motions that align with textual descriptions. These findings underscore the essential contributions of each component to the holistic success of the MLUG model, emphasizing that a delicate balance between motion and text understanding is crucial for optimizing performance in complex text-to-motion generation tasks.
5. Discussion
A core strength of MLUG is its ability to handle open-vocabulary inputs and generate emotionally expressive motions. The architecture’s use of a unimodal encoder with motion–text contrastive loss enables it to effectively align diverse textual descriptions with corresponding motion features. This allows the model to generalize well to new and unseen textual inputs, supporting its ability to handle open-vocabulary scenarios. Additionally, the integration of a motion-grounded text encoder further enhances the model’s capacity to understand and generate contextually appropriate and emotionally expressive motions. In our experiments, we observed that the model consistently generated motions that reflected emotional nuances described in the text, such as excitement, sadness, or calmness. These qualitative outcomes, along with quantitative metrics (e.g., diversity and precision scores), provide evidence of the model’s flexibility and effectiveness in producing semantically and emotionally rich motions.
While many of our quantitative metrics show improvements over state-of-the-art methods, we acknowledge that in some cases, the performance gains are incremental. However, it is important to note that MLUG consistently demonstrates superior performance in terms of motion diversity, semantic alignment, and emotionally expressive generation. To provide a more robust understanding of the significance of these improvements, we have now included confidence intervals for all evaluation metrics, particularly in the ablation studies. This additional statistical analysis highlights the consistency and reliability of our model’s performance, ensuring that even minor improvements are statistically significant and not due to random variations in the data. We believe these refinements offer a clearer and more accurate assessment of the model’s capabilities compared to previous works.
6. Related Work
Generating Motion Dynamics. Motion dynamics generation is diversified into several categories depending on the type of inputs used. For instance, some research has explored the use of music to drive the creation of dance movements [
1], while other studies have focused on generating motion from concise motion descriptions [
2,
3,
4] and specified action labels [
24,
25]. The effectiveness of these approaches largely relies on comprehensive motion capture databases [
5,
6,
7,
26,
27,
28] and databases labeled with motion descriptions, such as AMASS [
29], the KIT motion-language, and the HumanML3D database [
11]. Nonetheless, these databases are often constrained by their design and the challenges in data collection, including the omission of emotional movements. Despite achieving notable qualitative and quantitative results in some instances [
30,
31], methods trained on these limited datasets struggle to adapt to diverse motion descriptions.
Leveraging Pretrained Models for Knowledge Extraction. The advent of pretrained foundational models has unlocked the potential of zero-shot and few-shot learning, even outperforming traditional supervised learning methods in some cases [
9,
32,
33,
34]. Among these, the CLIP model [
9] stands out for its ability to align the semantic spaces of language and vision [
10]. When integrated with DALL-E, it offers impressive capabilities for generating images from textual descriptions. This remarkable representational capacity of foundational models has led to the development of zero-shot text-driven applications [
35,
36,
37,
38], including the generation of 3D meshes [
39,
40,
41,
42,
43].
Pre-training for Vision–Language Integration. Vision–language pre-training (VLP) seeks to enhance the performance on downstream vision and language tasks by initially training models on vast collections of image–text pairs. Given the high cost of obtaining human-annotated texts, most strategies [
44,
45,
46,
47] opt for web-crawled image and alt-text pairs [
48,
49,
50]. Despite employing basic rule-based filters to clean the data, significant noise remains in the collected web texts. This issue has been somewhat overshadowed by the improvements achieved through dataset scaling. This paper argues that noisy web texts are less than ideal for vision–language training and introduces CapFilt, a method to utilize web datasets more efficiently.
Various efforts have been made to consolidate different vision and language tasks within a singular framework [
47,
51,
52]. The principal challenge lies in designing model architectures capable of handling both understanding-based tasks (e.g., image–text retrieval) and generation-based tasks (e.g., image captioning). Models based solely on encoders [
46,
53] or encoder–decoders [
47,
52] have not been fully successful in excelling at both task types. However, our proposed model, a multimodal mixture of encoder–decoder, offers enhanced flexibility and superior performance across a broad spectrum of downstream tasks while maintaining a straightforward and efficient pre-training process.
7. Case Study
This case study showcases the top ten motion sequences retrieved from text-to-motion queries, highlighting our model’s ability to generalize and interpret various textual prompts. In one notable example, the model effectively interprets and responds to the concept of “strolling circularly”, a phrase it was not directly trained on. In another instance, the model successfully retrieves a set of motions that are in perfect harmony with the query “walking circularly”. The evaluation protocol identifies the motion at rank 1 as the most accurate, as it achieves a text similarity (TS) score above 0.95.
Figure 3 presents four scenarios where the model, initially trained on the Human3D dataset, is applied to sequences from the BABEL dataset. Each scenario displays the queried text at the top, with the frame count indicated horizontally. Motion features are computed using a rolling window approach, and the correlation between the text descriptor and a 20-frame window centered on each frame is depicted vertically. This is visualized over time as a 1D graph, demonstrating the model’s capability to anchor and align motion sequences with textual descriptions. It is also noteworthy that the model effectively bridges the domain gap between the BABEL labels used in testing and the Human3D dataset used during training.
8. Conclusions
MLUG stands as a significant leap forward in the domain of human motion generation, marking a departure from conventional methodologies to embrace the synergies of vision–language pre-training. This research elucidates the framework’s ability to seamlessly bridge the semantic gap between text and motion, facilitating the generation of movements that are not only physically plausible but also emotionally resonant and aligned with textual descriptions. Through rigorous evaluation across diverse datasets, MLUG has demonstrated superior performance over existing models, reflecting its robustness, versatility, and the high fidelity of generated motions. The success of MLUG underscores the potential of integrating advanced NLP techniques with motion synthesis to overcome historical challenges in the field, such as open-vocabulary generalization and the inclusion of emotional nuances. Looking ahead, MLUG not only sets a new standard for motion generation models but also offers a blueprint for future research aimed at further enriching virtual and interactive experiences.
Author Contributions
All authors contributed significantly to the conceptualization, methodology, and development of the MLUG framework. H.L. was responsible for the primary drafting of the manuscript and led the experimental design and data analysis. W.X. contributed to the model architecture and optimization strategy, offering key insights into the evaluation metrics and data processing pipeline. D.T. was instrumental in designing the visualization tools and contributed to refining the interpretation of the results. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Institutional Review Board Statement
This study did not involve any human participants or animal subjects and therefore did not require an Institutional Review Board (IRB) approval.
Informed Consent Statement
As this research did not involve human participants, informed consent was not applicable.
Data Availability Statement
The datasets used in this study, specifically the HumanML3D and KIT datasets, are publicly available and can be accessed through their respective sources as cited within the manuscript. Any additional information related to the data processing pipeline is available upon request from the corresponding author.
Conflicts of Interest
Author Daniel Tang was employed by the company “Mind Bridge AI, Ltd.”. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
References
- Aggarwal, G.; Parikh, D. Dance2Music: Automatic Dance-driven Music Generation. arXiv 2021, arXiv:2107.06252. [Google Scholar]
- Lin, X.; Amer, M.R. Human motion modeling using dvgans. arXiv 2018, arXiv:1804.10652. [Google Scholar]
- Ahuja, C.; Morency, L.P. Language2Pose: Natural Language Grounded Pose Forecasting. In Proceedings of the International Conference on 3D Vision, Quebec City, QC, Canada, 16–19 September 2019. [Google Scholar]
- Ahn, H.; Ha, T.; Choi, Y.; Yoo, H.; Oh, S. Text2Action: Generative Adversarial Synthesis from Language to Action. In Proceedings of the International Conference on Robotics and Automation, Brisbane, Australia, 21–25 May 2018. [Google Scholar]
- Cai, Z.; Ren, D.; Zeng, A.; Lin, Z.; Yu, T.; Wang, W.; Fan, X.; Gao, Y.; Yu, Y.; Pan, L.; et al. HuMMan: Multi-Modal 4D Human Dataset for Versatile Sensing and Modeling. arXiv 2022, arXiv:2204.13686. [Google Scholar]
- Cai, Z.; Zhang, M.; Ren, J.; Wei, C.; Ren, D.; Li, J.; Lin, Z.; Zhao, H.; Yi, S.; Yang, L.; et al. Playing for 3D human recovery. arXiv 2021, arXiv:2110.07588. [Google Scholar] [CrossRef] [PubMed]
- Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1325–1339. [Google Scholar] [CrossRef] [PubMed]
- Memar Ardestani, M.; Yan, H. Noise reduction in human motion-captured signals for computer animation based on B-spline filtering. Sensors 2022, 22, 4629. [Google Scholar] [CrossRef] [PubMed]
- Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning Transferable Visual Models From Natural Language Supervision. In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021. [Google Scholar]
- Vinker, Y.; Pajouheshgar, E.; Bo, J.Y.; Bachmann, R.C.; Bermano, A.H.; Cohen-Or, D.; Zamir, A.; Shamir, A. CLIPasso: Semantically-Aware Object Sketching. ACM Trans. Graph. 2022, 41, 86. [Google Scholar] [CrossRef]
- Guo, C.; Zou, S.; Zuo, X.; Wang, S.; Ji, W.; Li, X.; Cheng, L. Generating Diverse and Natural 3D Human Motions From Text. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5152–5161. [Google Scholar]
- Plappert, M.; Mandery, C.; Asfour, T. The KIT motion-language dataset. Big Data 2016, 4, 236–252. [Google Scholar] [CrossRef]
- Mahmood, N.; Ghorbani, N.; Troje, N.F.; Pons-Moll, G.; Black, M.J. Amass: Archive of motion capture as surface shapes. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
- Chen, X.; Jiang, B.; Liu, W.; Huang, Z.; Fu, B.; Chen, T.; Yu, J.; Yu, G. Executing your commands via motion diffusion in latent space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
- Tevet, G.; Raab, S.; Gordon, B.; Shafir, Y.; Bermano, A.H.; Cohen-Or, D. Human motion diffusion model. arXiv 2022, arXiv:2209.14916. [Google Scholar]
- Yuan, Y.; Kitani, K. Dlow: Diversifying latent flows for diverse human motion prediction. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part IX 16; Springer: Berlin/Heidelberg, Germany, 2020; pp. 346–364. [Google Scholar]
- Zhang, Y.; Black, M.J.; Tang, S. We are more than our joints: Predicting how 3d bodies move. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3372–3382. [Google Scholar]
- Ma, H.; Li, J.; Hosseini, R.; Tomizuka, M.; Choi, C. Multi-objective diverse human motion prediction with knowledge distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8161–8171. [Google Scholar]
- Guo, C.; Zuo, X.; Wang, S.; Cheng, L. Tm2t: Stochastic and tokenized modeling for the reciprocal generation of 3d human motions and texts. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
- Papineni, K.; Roukos, S.; Ward, T.; Zhu, W.J. Bleu: A method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, Philadelphia, PA, USA, 7–12 July 2002; pp. 311–318. [Google Scholar]
- Lin, C.Y. Rouge: A package for automatic evaluation of summaries. In Proceedings of the Text Summarization Branches Out, Barcelona, Spain, 25–26 July 2004; pp. 74–81. [Google Scholar]
- Vedantam, R.; Zitnick, C.L.; Parikh, D. Cider: Consensus-based image description evaluation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4566–4575. [Google Scholar]
- Zhang, T.; Kishore, V.; Wu, F.; Weinberger, K.Q.; Artzi, Y. Bertscore: Evaluating text generation with bert. arXiv 2019, arXiv:1904.09675. [Google Scholar]
- Guo, C.; Zuo, X.; Wang, S.; Zou, S.; Sun, Q.; Deng, A.; Gong, M.; Cheung, L. Action2motion: Conditioned generation of 3d human motions. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; ACM: New York, NY, USA, 2020; pp. 2021–2029. [Google Scholar]
- Petrovich, M.; Black, M.J.; Varol, G. Action-Conditioned 3D Human Motion Synthesis with Transformer VAE. In Proceedings of the International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
- Mehta, D.; Rhodin, H.; Casas, D.; Fua, P.; Sotnychenko, O.; Xu, W.; Theobalt, C. Monocular 3D Human Pose Estimation In The Wild Using Improved CNN Supervision. In Proceedings of the International Conference on 3D Vision, Stanford, CA, USA, 25–28 October 2016. [Google Scholar]
- Varol, G.; Romero, J.; Martin, X.; Mahmood, N.; Black, M.J.; Laptev, I.; Schmid, C. Learning from Synthetic Humans. In Proceedings of the Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
- von Marcard, T.; Henschel, R.; Black, M.J.; Rosenhahn, B.; Pons-Moll, G. Recovering Accurate {3D} Human Pose in the Wild Using {IMUs} and a Moving Camera. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar]
- Punnakkal, A.R.; Chandrasekaran, A.; Athanasiou, N.; Quiros-Ramirez, A.; Black, M.J. BABEL: Bodies, Action and Behavior with English Labels. In Proceedings of the Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
- Athanasiou, N.; Petrovich, M.; Black, M.J.; Varol, G. TEACH: Temporal Action Compositions for 3D Humans. In Proceedings of the International Conference on 3D Vision (3DV), Prague, Czech Republic, 12–16 September 2022. [Google Scholar]
- Zhang, M.; Cai, Z.; Pan, L.; Hong, F.; Guo, X.; Yang, L.; Liu, Z. MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model. arXiv 2022, arXiv:2208.15001. [Google Scholar] [CrossRef] [PubMed]
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
- Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
- Wang, L.; Gong, Y.; Ma, X.; Wang, Q.; Zhou, K.; Chen, L. IS-MVSNet: Importance Sampling-Based MVSNet. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 668–683. [Google Scholar]
- Frans, K.; Soros, L.B.; Witkowski, O. CLIPDraw: Exploring Text-to-Drawing Synthesis through Language-Image Encoders. arXiv 2021, arXiv:2106.14843. [Google Scholar]
- Patashnik, O.; Wu, Z.; Shechtman, E.; Cohen-Or, D.; Lischinski, D. StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery. arXiv 2021, arXiv:2103.17249. [Google Scholar]
- Peng, S.; Zhang, Y.; Xu, Y.; Wang, Q.; Shuai, Q.; Bao, H.; Zhou, X. Neural body: Implicit neural representations with structured latent codes for novel view synthesis of dynamic humans. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9054–9063. [Google Scholar]
- Huang, R.; Zhong, W.; Li, G. Audio-Driven Talking Head Generation with Transformer and 3D Morphable Model. In Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022; MM ’22. pp. 7035–7039. [Google Scholar] [CrossRef]
- Jain, A.; Mildenhall, B.; Barron, J.T.; Abbeel, P.; Poole, B. Zero-shot text-guided object generation with dream fields. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2022), New Orleans, LA, USA, 18–24 June 2022; pp. 867–876. [Google Scholar]
- Jetchev, N. ClipMatrix: Text-controlled Creation of 3D Textured Meshes. arXiv 2022, arXiv:2109.12922. [Google Scholar]
- Michel, O.; Bar-On, R.; Liu, R.; Benaim, S.; Hanocka, R. Text2Mesh: Text-Driven Neural Stylization for Meshes. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
- Sanghi, A.; Chu, H.; Lambourne, J.G.; Wang, Y.; Cheng, C.Y.; Fumero, M. CLIP-Forge: Towards Zero-Shot Text-to-Shape Generation. arXiv 2021, arXiv:2110.02624. [Google Scholar]
- Peng, S.; Dong, J.; Wang, Q.; Zhang, S.; Shuai, Q.; Bao, H.; Zhou, X. Animatable neural radiance fields for human body modeling. arXiv 2021, arXiv:2105.02872. [Google Scholar]
- Chen, Y.; Li, L.; Yu, L.; Kholy, A.E.; Ahmed, F.; Gan, Z.; Cheng, Y.; Liu, J. UNITER: UNiversal Image-TExt Representation Learning. In Proceedings of the ECCV, Glasgow, UK, 23–28 August 2020; Volume 12375, pp. 104–120. [Google Scholar]
- Li, X.; Yin, X.; Li, C.; Zhang, P.; Hu, X.; Zhang, L.; Wang, L.; Hu, H.; Dong, L.; Wei, F.; et al. Oscar: Object-Semantics Aligned Pre-training for Vision-Language Tasks. In Proceedings of the ECCV, Glasgow, UK, 23–28 August 2020; pp. 121–137. [Google Scholar]
- Li, J.; Selvaraju, R.R.; Gotmare, A.D.; Joty, S.; Xiong, C.; Hoi, S. Align before Fuse: Vision and Language Representation Learning with Momentum Distillation. In Proceedings of the 35th International Conference on Neural Information Processing Systems, Online, 6–14 December 2021. [Google Scholar]
- Wang, Z.; Yu, J.; Yu, A.W.; Dai, Z.; Tsvetkov, Y.; Cao, Y. SimVLM: Simple visual language model pretraining with weak supervision. arXiv 2021, arXiv:2108.10904. [Google Scholar]
- Sharma, P.; Ding, N.; Goodman, S.; Soricut, R. Conceptual Captions: A Cleaned, Hypernymed, Image Alt-text Dataset For Automatic Image Captioning. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Long Papers), Melbourne, Australia, 15–20 July 2018; Gurevych, I., Miyao, Y., Eds.; pp. 2556–2565. [Google Scholar]
- Changpinyo, S.; Sharma, P.; Ding, N.; Soricut, R. Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2021), Virtual Conference, 19–25 June 2021. [Google Scholar]
- Jia, C.; Yang, Y.; Xia, Y.; Chen, Y.T.; Parekh, Z.; Pham, H.; Le, Q.V.; Sung, Y.; Li, Z.; Duerig, T. Scaling up visual and vision-language representation learning with noisy text supervision. arXiv 2021, arXiv:2102.05918. [Google Scholar]
- Zhou, L.; Palangi, H.; Zhang, L.; Hu, H.; Corso, J.J.; Gao, J. Unified Vision-Language Pre-Training for Image Captioning and VQA. In Proceedings of the 34th AAAI Conference on Artificial Intelligence (AAAI-20), New York, NY, USA, 7–12 February 2020; pp. 13041–13049. [Google Scholar]
- Cho, J.; Lei, J.; Tan, H.; Bansal, M. Unifying vision-and-language tasks via text generation. arXiv 2021, arXiv:2102.02779. [Google Scholar]
- Li, W.; Gao, C.; Niu, G.; Xiao, X.; Liu, H.; Liu, J.; Wu, H.; Wang, H. UNIMO: Towards Unified-Modal Understanding and Generation via Cross-Modal Contrastive Learning. In Proceedings of the ACL-IJCNLP 2021 Student Research Workshop, ACL 2021, Online, 5–10 July 2021; Zong, C., Xia, F., Li, W., Navigli, R., Eds.; pp. 2592–2607. [Google Scholar]
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).