add elevation in call
Browse files- data/corgi.jpg +0 -0
- pipeline_mvdream.py +3 -4
- run_imagedream.py +1 -1
- run_mvdream.py +1 -1
    	
        data/corgi.jpg
    ADDED
    
    |   | 
    	
        pipeline_mvdream.py
    CHANGED
    
    | @@ -435,6 +435,7 @@ class MVDreamPipeline(DiffusionPipeline): | |
| 435 | 
             
                    image: Optional[np.ndarray] = None,
         | 
| 436 | 
             
                    height: int = 256,
         | 
| 437 | 
             
                    width: int = 256,
         | 
|  | |
| 438 | 
             
                    num_inference_steps: int = 50,
         | 
| 439 | 
             
                    guidance_scale: float = 7.0,
         | 
| 440 | 
             
                    negative_prompt: str = "",
         | 
| @@ -489,10 +490,8 @@ class MVDreamPipeline(DiffusionPipeline): | |
| 489 | 
             
                        None,
         | 
| 490 | 
             
                    )
         | 
| 491 |  | 
| 492 | 
            -
                     | 
| 493 | 
            -
             | 
| 494 | 
            -
                    else:
         | 
| 495 | 
            -
                        camera = get_camera(num_frames, elevation=15, extra_view=False).to(dtype=latents.dtype, device=device)
         | 
| 496 | 
             
                    camera = camera.repeat_interleave(num_images_per_prompt, dim=0)
         | 
| 497 |  | 
| 498 | 
             
                    # Prepare extra step kwargs.
         | 
|  | |
| 435 | 
             
                    image: Optional[np.ndarray] = None,
         | 
| 436 | 
             
                    height: int = 256,
         | 
| 437 | 
             
                    width: int = 256,
         | 
| 438 | 
            +
                    elevation: float = 0,
         | 
| 439 | 
             
                    num_inference_steps: int = 50,
         | 
| 440 | 
             
                    guidance_scale: float = 7.0,
         | 
| 441 | 
             
                    negative_prompt: str = "",
         | 
|  | |
| 490 | 
             
                        None,
         | 
| 491 | 
             
                    )
         | 
| 492 |  | 
| 493 | 
            +
                    # Get camera
         | 
| 494 | 
            +
                    camera = get_camera(num_frames, elevation=elevation, extra_view=(image is not None)).to(dtype=latents.dtype, device=device)
         | 
|  | |
|  | |
| 495 | 
             
                    camera = camera.repeat_interleave(num_images_per_prompt, dim=0)
         | 
| 496 |  | 
| 497 | 
             
                    # Prepare extra step kwargs.
         | 
    	
        run_imagedream.py
    CHANGED
    
    | @@ -20,7 +20,7 @@ args = parser.parse_args() | |
| 20 |  | 
| 21 | 
             
            for i in range(5):
         | 
| 22 | 
             
                input_image = kiui.read_image(args.image, mode='float')
         | 
| 23 | 
            -
                image = pipe(args.prompt, input_image, guidance_scale=5)
         | 
| 24 | 
             
                grid = np.concatenate(
         | 
| 25 | 
             
                    [
         | 
| 26 | 
             
                        np.concatenate([image[0], image[2]], axis=0),
         | 
|  | |
| 20 |  | 
| 21 | 
             
            for i in range(5):
         | 
| 22 | 
             
                input_image = kiui.read_image(args.image, mode='float')
         | 
| 23 | 
            +
                image = pipe(args.prompt, input_image, guidance_scale=5, num_inference_steps=30, elevation=0)
         | 
| 24 | 
             
                grid = np.concatenate(
         | 
| 25 | 
             
                    [
         | 
| 26 | 
             
                        np.concatenate([image[0], image[2]], axis=0),
         | 
    	
        run_mvdream.py
    CHANGED
    
    | @@ -19,7 +19,7 @@ parser.add_argument("prompt", type=str, default="a cute owl 3d model") | |
| 19 | 
             
            args = parser.parse_args()
         | 
| 20 |  | 
| 21 | 
             
            for i in range(5):
         | 
| 22 | 
            -
                image = pipe(args.prompt)
         | 
| 23 | 
             
                grid = np.concatenate(
         | 
| 24 | 
             
                    [
         | 
| 25 | 
             
                        np.concatenate([image[0], image[2]], axis=0),
         | 
|  | |
| 19 | 
             
            args = parser.parse_args()
         | 
| 20 |  | 
| 21 | 
             
            for i in range(5):
         | 
| 22 | 
            +
                image = pipe(args.prompt, guidance_scale=5, num_inference_steps=30, elevation=0)
         | 
| 23 | 
             
                grid = np.concatenate(
         | 
| 24 | 
             
                    [
         | 
| 25 | 
             
                        np.concatenate([image[0], image[2]], axis=0),
         | 

