Spaces:
Running
Running
changed the examples
Browse files- app.py +4 -8
- example_images/.DS_Store +0 -0
- example_images/aaron.jpeg +0 -0
- example_images/demo-1.jpg +0 -0
- example_images/demo-2.jpg +0 -0
- example_images/docvqa_example.png +0 -3
- example_images/examples_invoice.png +0 -0
- example_images/examples_weather_events.png +0 -0
- example_images/rococo_1.jpg +0 -0
- example_images/s2w_example.png +0 -0
app.py
CHANGED
|
@@ -36,16 +36,13 @@ def model_inference(input_dict, history):
|
|
| 36 |
|
| 37 |
|
| 38 |
examples=[
|
|
|
|
|
|
|
| 39 |
[{"text": "What art era do this artpiece belong to?", "files": ["example_images/rococo.jpg"]}, []],
|
| 40 |
[{"text": "Caption", "files": ["example_images/rococo.jpg"]}, []],
|
| 41 |
[{"text": "I'm planning a visit to this temple, give me travel tips.", "files": ["example_images/examples_wat_arun.jpg"]}, []],
|
| 42 |
[{"text": "Caption", "files": ["example_images/examples_wat_arun.jpg"]}, []],
|
| 43 |
-
[{"text": "
|
| 44 |
-
[{"text": "Caption", "files": ["example_images/examples_invoice.png"]}, []],
|
| 45 |
-
[{"text": "What is this UI about?", "files": ["example_images/s2w_example.png"]}, []],
|
| 46 |
-
[{"text": "Caption", "files": ["example_images/s2w_example.png"]}, []],
|
| 47 |
-
[{"text": "Where do the severe droughts happen according to this diagram?", "files": ["example_images/examples_weather_events.png"]}, []],
|
| 48 |
-
[{"text": "Caption", "files": ["example_images/examples_weather_events.png"]}, []],
|
| 49 |
]
|
| 50 |
|
| 51 |
demo = gr.ChatInterface(fn=model_inference, title="Moondream 0.5B: The World's Smallest Vision-Language Model",
|
|
@@ -54,5 +51,4 @@ demo = gr.ChatInterface(fn=model_inference, title="Moondream 0.5B: The World's S
|
|
| 54 |
textbox=gr.MultimodalTextbox(label="Query Input", file_types=["image"], file_count="single"), stop_btn="Stop Generation", multimodal=True,
|
| 55 |
additional_inputs=[], cache_examples=False)
|
| 56 |
|
| 57 |
-
demo.launch(debug=True)
|
| 58 |
-
|
|
|
|
| 36 |
|
| 37 |
|
| 38 |
examples=[
|
| 39 |
+
[{"text": "Caption", "files": ["example_images/demo-1.jpg"]}, []],
|
| 40 |
+
[{"text": "Caption", "files": ["example_images/demo-2.jpg"]}, []],
|
| 41 |
[{"text": "What art era do this artpiece belong to?", "files": ["example_images/rococo.jpg"]}, []],
|
| 42 |
[{"text": "Caption", "files": ["example_images/rococo.jpg"]}, []],
|
| 43 |
[{"text": "I'm planning a visit to this temple, give me travel tips.", "files": ["example_images/examples_wat_arun.jpg"]}, []],
|
| 44 |
[{"text": "Caption", "files": ["example_images/examples_wat_arun.jpg"]}, []],
|
| 45 |
+
[{"text": "Caption", "files": ["example_images/aaron.jpeg"]}, []],
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
]
|
| 47 |
|
| 48 |
demo = gr.ChatInterface(fn=model_inference, title="Moondream 0.5B: The World's Smallest Vision-Language Model",
|
|
|
|
| 51 |
textbox=gr.MultimodalTextbox(label="Query Input", file_types=["image"], file_count="single"), stop_btn="Stop Generation", multimodal=True,
|
| 52 |
additional_inputs=[], cache_examples=False)
|
| 53 |
|
| 54 |
+
demo.launch(debug=True)
|
|
|
example_images/.DS_Store
ADDED
|
Binary file (6.15 kB). View file
|
|
|
example_images/aaron.jpeg
ADDED
|
example_images/demo-1.jpg
ADDED
|
example_images/demo-2.jpg
ADDED
|
example_images/docvqa_example.png
DELETED
Git LFS Details
|
example_images/examples_invoice.png
DELETED
|
Binary file (50 kB)
|
|
|
example_images/examples_weather_events.png
DELETED
|
Binary file (310 kB)
|
|
|
example_images/rococo_1.jpg
DELETED
|
Binary file (849 kB)
|
|
|
example_images/s2w_example.png
DELETED
|
Binary file (82.8 kB)
|
|
|