Abstract

Currently, generating content tied to specific locations via genAI models can often result in hallucinations of non-existent places or creation of generic content that is devoid of the target specifics. The difficulties in generating content for specific locations result from the genAI models being trained on massive datasets containing images from many locations without sufficient focus on a single location. This disclosure describes techniques to enhance the creation of location-specific personalized 2D or 3D immersive or non-immersive multimedia experiences (e.g., 2D or 3D panoramic video with or without sound) via multimodal genAI models. Base genAI models are boosted with multimedia data integrated from multiple diverse sources to provide the model with context for a specific location. Models boosted in this manner can be used to generate 2D or 3D multimedia experiences that are customized for a specific location or context. Additionally, with user permission, model outputs can be personalized according to user preferences.

Creative Commons License

Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.

Share

COinS