Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Define inference configs for more models #1069

Merged
merged 13 commits into from
Jan 11, 2025
Merged

Define inference configs for more models #1069

merged 13 commits into from
Jan 11, 2025

Conversation

nikg4
Copy link
Collaborator

@nikg4 nikg4 commented Jan 11, 2025

Description

-- Define inference configs for more VLM models, and add to README
-- Add missing entry for Llama 1B in README
-- Update docs to show how to use oumi infer for multimodal

Related issues

Towards OPE-733, OPE-862

Before submitting

  • This PR only changes documentation. (You can ignore the following checks in that case)
  • Did you read the contributor guideline Pull Request guidelines?
  • Did you link the issue(s) related to this PR in the section above?
  • Did you add / update tests where needed?

Reviewers

At least one review from a member of oumi-ai/oumi-staff is required.

Copy link

linear bot commented Jan 11, 2025

OPE-733

OPE-862

@nikg4 nikg4 changed the title Define inference configs for model models Define inference configs for more models Jan 11, 2025
@nikg4 nikg4 marked this pull request as ready for review January 11, 2025 02:10
To run multimodal inference interactively, use the `oumi infer` command with the `-i` and `--image` flags.

```{code-block} bash
oumi infer -c infer_config.yaml -i --image="https://oumi.ai/favicon.ico"
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We should find a better image to use as an example... @taenin maybe we can host something on the website ?

batch_size: 1

# Sample command to start SGLang server:
# CUDA_VISIBLE_DEVICES=0 python -m sglang.launch_server \
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

nit: maybe we can put command example at the start of the file?

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

updated

@nikg4 nikg4 merged commit 4831733 into main Jan 11, 2025
1 check passed
@nikg4 nikg4 deleted the xrdaukar/vlm-infer-cfg branch January 11, 2025 20:59
@nikg4 nikg4 mentioned this pull request Jan 11, 2025
4 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants