Skip to content
This repository has been archived by the owner on Nov 27, 2024. It is now read-only.

TextGeneration Pipeline #131

Open
wants to merge 2 commits into
base: master
Choose a base branch
from
Open

TextGeneration Pipeline #131

wants to merge 2 commits into from

Conversation

saddam213
Copy link
Member

@saddam213 saddam213 commented Mar 20, 2024

Support for TextGeneration ONNX models

Initial support will be using OnnxRuntime-GenAI onnxruntime-genai

TODO:

  • TextGeneration Project
  • Basic stateless pipeline
  • CUDA and CPU support
  • Code Examples

@AshD
Copy link

AshD commented Apr 5, 2024

Looks cool. How is the text generation speed for onnx models compared to llamasharp for a Mistral 7B v0.2 Instruct model for example?

@saddam213
Copy link
Member Author

saddam213 commented Apr 5, 2024

Looks cool. How is the text generation speed for onnx models compared to llamasharp for a Mistral 7B v0.2 Instruct model for example?

Im not too sure, I have not been able to get LLamaSharp working in Web or WPF since version 0.8. so this is my first attempt using another LLM base library, this one is still very new but its a million times less complicated than llama.cpp

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants