Skip to content

Basic Llama 2 and model customization support#3

Open
athewsey wants to merge 2 commits intodalacan:mainfrom
athewsey:feat/misc
Open

Basic Llama 2 and model customization support#3
athewsey wants to merge 2 commits intodalacan:mainfrom
athewsey:feat/misc

Conversation

@athewsey
Copy link
Contributor

  1. Pass through CustomAttributes header in Falcon and Flan Lambda functions, and add an example config for using Llama 2 (which requires a CustomAttribute for EULA approval to be sent with every request) with the Falcon function
    • I'm not sure if direct API integration is also possible with Llama? Didn't have time to test
  2. Support overriding model tarball location from config
    • We used this to run a Falcon endpoint with modified inference handler

Appreciate more formal/documented Llama 2 support might be preferable for (1), and (2) might be a bit niche... But was low on time so just offering what we have :-)

Extend Falcon and Flan Lambda functions to support passing through
CustomAttributes header. With this change the Falcon Lambda should be
usable with Llama 2 model on SMJS - which requires an accept_eula
CustomAttribute on every call.
Add and document (in README) options to override model tar location
from config file. We used this to customise the inference handler for
a deployed model.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant