peteromallet commited on
Commit
a0d4cff
·
verified ·
1 Parent(s): a93b0e9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -13
README.md CHANGED
@@ -1,7 +1,5 @@
1
  ---
2
  license: apache-2.0
3
- datasets:
4
- - peteromallet/high-quality-midjouney-srefs
5
  base_model:
6
  - Qwen/Qwen-Image-Edit
7
  tags:
@@ -18,7 +16,7 @@ library_name: diffusers
18
 
19
  ## Model Description
20
 
21
- **InScene** and **InScene Annotate** are a pair of LoRA fine-tunes for QwenEdit that enhance its ability to generate images based on scene references. These models work together to provide flexible scene-based image generation with optional annotation support.
22
 
23
  ### InScene
24
  The main model that generates images based on scene composition and layout from a reference image. InScene is trained on pairs of different shots within the same scene, along with prompts describing the desired output. Its goal is to create entirely new shots within a scene while maintaining character consistency and scene coherence.
@@ -39,20 +37,21 @@ InScene Annotate is trained on images with green rectangles drawn over specific
39
  ### InScene
40
  To use the base InScene model, start your prompt with:
41
 
42
- `Make an image in this scene of `
43
 
44
  And then describe what you want to generate.
45
 
46
  For example:
47
- `Make an image in this scene of a bustling city street at night.`
48
 
49
  ### InScene Annotate
50
- For the annotate variant, use annotated reference images and start your prompt with:
51
 
52
- `Based on this annotated scene, create `
 
53
 
54
  For example:
55
- `Based on this annotated scene, create a winter landscape with snow-covered mountains.`
56
 
57
  ### Use with diffusers
58
 
@@ -95,12 +94,10 @@ The models may struggle with:
95
 
96
  ## Training Data
97
 
98
- The InScene and InScene Annotate LoRAs were trained on a curated dataset of high-quality Midjourney style references, with additional scene-focused annotations for the Annotate variant. The dataset emphasizes diverse scene compositions and spatial relationships.
99
 
100
- You can find the public dataset used for training here:
101
- [https://huggingface.co/datasets/peteromallet/high-quality-midjouney-srefs](https://huggingface.co/datasets/peteromallet/high-quality-midjouney-srefs)
102
 
103
  ## Links
104
 
105
- - Model: [https://huggingface.co/peteromallet/Qwen-Image-Edit-InScene](https://huggingface.co/peteromallet/Qwen-Image-Edit-InScene)
106
- - Dataset: [https://huggingface.co/datasets/peteromallet/high-quality-midjouney-srefs](https://huggingface.co/datasets/peteromallet/high-quality-midjouney-srefs)
 
1
  ---
2
  license: apache-2.0
 
 
3
  base_model:
4
  - Qwen/Qwen-Image-Edit
5
  tags:
 
16
 
17
  ## Model Description
18
 
19
+ **InScene** and **InScene Annotate** are a pair of LoRA fine-tunes for QwenEdit that enhance its ability to generate images based on scene references. These models work together to provide flexible scene-based image generation with optional annotation support. **Both models are currently in beta and will be improved significantly over time.**
20
 
21
  ### InScene
22
  The main model that generates images based on scene composition and layout from a reference image. InScene is trained on pairs of different shots within the same scene, along with prompts describing the desired output. Its goal is to create entirely new shots within a scene while maintaining character consistency and scene coherence.
 
37
  ### InScene
38
  To use the base InScene model, start your prompt with:
39
 
40
+ `Show a different image in the same scene of: `
41
 
42
  And then describe what you want to generate.
43
 
44
  For example:
45
+ `Show a different image in the same scene of: a bustling city street at night.`
46
 
47
  ### InScene Annotate
48
+ To use InScene Annotate:
49
 
50
+ 1. Draw a green rectangle over the subject or area of interest in your reference image
51
+ 2. Describe what you want to focus on and how it should change
52
 
53
  For example:
54
+ `Zoom in on the girl, make her turn to the side and laugh`
55
 
56
  ### Use with diffusers
57
 
 
94
 
95
  ## Training Data
96
 
97
+ The InScene and InScene Annotate LoRAs were trained on curated datasets focusing on scene composition and spatial relationships. InScene uses pairs of different shots within the same scene, while InScene Annotate uses annotated images with green rectangle markers.
98
 
99
+ The training data will be released publicly when it's in a more stable state.
 
100
 
101
  ## Links
102
 
103
+ - Model: [https://huggingface.co/peteromallet/Qwen-Image-Edit-InScene](https://huggingface.co/peteromallet/Qwen-Image-Edit-InScene)