-
Notifications
You must be signed in to change notification settings - Fork 547
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Define inference configs for more models #1069
Conversation
To run multimodal inference interactively, use the `oumi infer` command with the `-i` and `--image` flags. | ||
|
||
```{code-block} bash | ||
oumi infer -c infer_config.yaml -i --image="https://oumi.ai/favicon.ico" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We should find a better image to use as an example... @taenin maybe we can host something on the website ?
batch_size: 1 | ||
|
||
# Sample command to start SGLang server: | ||
# CUDA_VISIBLE_DEVICES=0 python -m sglang.launch_server \ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nit: maybe we can put command example at the start of the file?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
updated
Description
-- Define inference configs for more VLM models, and add to README
-- Add missing entry for Llama 1B in README
-- Update docs to show how to use
oumi infer
for multimodalRelated issues
Towards OPE-733, OPE-862
Before submitting
Reviewers
At least one review from a member of
oumi-ai/oumi-staff
is required.