Skip to main content

SKED: sketch-guided text-based 3D editing

Resource type
Thesis type
(Thesis) M.Sc.
Date created
2024-05-31
Authors/Contributors
Abstract
With the recent advent of Text-to-3D models based on Text-to-image diffusion models, they are increasingly being integrated into Computer Graphics pipelines, facilitating creative 3D design tasks in unrestricted domains. Nevertheless, simple textual prompting is often insufficient for interactive and local manipulation of 3D content. Incorporating user-guided sketches offers a more intuitive control for editing and manipulation. However, since state-of-the-art Text-to-3D models rely on gradients generated by rendering arbitrary views for optimization, integrating sketches in these pipelines is not straightforward. This thesis presents SKED, a method for editing 3D objects represented as Neural Radiance Fields (NeRFs) conditioned on a text prompt and two sketches from different views. We ensure semantic alignment of the edit by using a pretrained text-to-image model as guidance. Additionally, our framework uses novel loss functions to make the edit adhere to the provided sketches. We demonstrate the effectiveness of our method through various qualitative and quantitative experiments.
Document
Extent
34 pages.
Identifier
etd23105
Copyright statement
Copyright is held by the author(s).
Permissions
This thesis may be printed or downloaded for non-commercial research and scholarly purposes.
Supervisor or Senior Supervisor
Thesis advisor: Mahdavi-Amiri, Ali
Language
English
Member of collection
Download file Size
etd23105.pdf 9.34 MB

Views & downloads - as of June 2023

Views: 0
Downloads: 0