Commit graph

  • 58c6001be9 Add missing exllamav2 samplers oobabooga 2023-11-16 07:09:40 -08:00
  • cd41f8912b Warn users about n_ctx / max_seq_len oobabooga 2023-11-15 18:56:42 -08:00
  • a475aa7816 Improve API documentation oobabooga 2023-11-15 18:39:08 -08:00
  • 9be48e83a9 Start API when "api" checkbox is checked oobabooga 2023-11-15 16:35:47 -08:00
  • a85ce5f055 Add more info messages for truncation / instruction template oobabooga 2023-11-15 16:13:36 -08:00
  • 883701bc40 Alternative solution to 025da386a0 oobabooga 2023-11-15 16:04:02 -08:00
  • 8ac942813c Revert "Fix CPU memory limit error (issue #3763) (#4597)" oobabooga 2023-11-15 16:01:54 -08:00
  • e6f44d6d19 Print context length / instruction template to terminal when loading models oobabooga 2023-11-15 16:00:51 -08:00
  • e05d8fd441 Style changes oobabooga 2023-11-15 15:51:37 -08:00
  • be125e2708 Add /v1/internal/model/unload endpoint oobabooga 2023-11-15 15:48:33 -08:00
  • 564d0cde82
    Use standard hyphens in filenames (#4576) David Nielson 2023-11-15 16:29:00 -07:00
  • 025da386a0
    Fix CPU memory limit error (issue #3763) (#4597) Andy Bao 2023-11-15 18:27:20 -05:00
  • 8a9d5a0cea
    update AutoGPTQ to higher version for lora applying error fixing (#4604) Anton Rogozin 2023-11-16 04:23:22 +05:00
  • 8a2af87d3a
    Merge pull request #4608 from oobabooga/dev oobabooga 2023-11-15 13:19:15 -03:00
  • 072cfe19e9 Minor Colab fix oobabooga 2023-11-15 08:18:32 -08:00
  • 2337aebe4d
    Merge pull request #4606 from oobabooga/dev oobabooga 2023-11-15 13:16:44 -03:00
  • 3d861a459d Minor Colab fix oobabooga 2023-11-15 08:15:43 -08:00
  • dea90c7b67 Bump exllamav2 to 0.0.8 oobabooga 2023-11-13 10:34:10 -08:00
  • 454fcf39a9
    Merge pull request #4579 from oobabooga/dev oobabooga 2023-11-13 11:39:08 -03:00
  • 4f9bc63edf Installer: update a message for clarity oobabooga 2023-11-10 09:43:02 -08:00
  • 74fee4f312 Update Colab-TextGen-GPU.ipynb oobabooga 2023-11-10 09:18:25 -08:00
  • 52758f15da Remove sentence-transformers requirement (for #1575) oobabooga 2023-11-10 07:35:29 -08:00
  • c5be3f7acb Make /v1/embeddings functional, add request/response types oobabooga 2023-11-10 07:34:27 -08:00
  • 7ed2143cd6
    Update 12 - OpenAI API.md oobabooga 2023-11-10 11:56:04 -03:00
  • 0777b0d3c7 Add system_message parameter, document model (unused) parameter oobabooga 2023-11-10 06:47:00 -08:00
  • 4aabff3728 Remove old API, launch OpenAI API with --api oobabooga 2023-11-10 06:39:08 -08:00
  • 6a7cd01ebf
    Fix bug with /internal/model/load (#4549) GuizzyQC 2023-11-09 22:16:38 -05:00
  • 2af7e382b1 Revert "Bump llama-cpp-python to 0.2.14" oobabooga 2023-11-09 10:00:22 -08:00
  • 07d66e45b4
    Merge pull request #4541 from oobabooga/dev oobabooga 2023-11-09 14:53:34 -03:00
  • 372d712921
    Fix deprecated API (#4539) Ashley Kleynhans 2023-11-09 19:51:50 +02:00
  • d86f1fd2c3 OpenAI API: stop streaming on client disconnect (closes #4521) oobabooga 2023-11-09 06:37:32 -08:00
  • f7534b2f4b
    Merge pull request #4532 from oobabooga/dev oobabooga 2023-11-09 09:33:55 -03:00
  • effb3aef42 Prevent deadlocks in OpenAI API with simultaneous requests oobabooga 2023-11-08 20:55:39 -08:00
  • 4da00b6032
    Merge pull request #4522 from oobabooga/dev oobabooga 2023-11-08 22:57:08 -03:00
  • 21ed9a260e Document the new "Custom system message" field oobabooga 2023-11-08 17:54:10 -08:00
  • 678fd73aef Document /v1/internal/model/load and fix a bug oobabooga 2023-11-08 17:41:12 -08:00
  • 1754a3761b
    Include trust remote code usage in openai api's embedder (#4513) MrMojoR 2023-11-08 15:25:43 +01:00
  • 6c7aad11f3
    openai extension: wrong frequency_penalty type (#4512) hronoas 2023-11-08 17:23:51 +03:00
  • 881e8a6e70
    Small bug fix in /v1/internal/model/load oobabooga 2023-11-08 02:34:13 -03:00
  • 050ff36bd6 Revert "Add a comment to /v1/models" oobabooga 2023-11-07 21:09:47 -08:00
  • 38b07493a0 Add a comment to /v1/models oobabooga 2023-11-07 21:07:12 -08:00
  • 2358706453 Add /v1/internal/model/load endpoint (tentative) oobabooga 2023-11-07 20:58:06 -08:00
  • 43c53a7820 Refactor the /v1/models endpoint oobabooga 2023-11-07 19:59:27 -08:00
  • 1b69694fe9 Add types to the encode/decode/token-count endpoints oobabooga 2023-11-07 19:05:36 -08:00
  • f6ca9cfcdc Add /v1/internal/model-info endpoint oobabooga 2023-11-07 18:59:02 -08:00
  • 6e2e0317af
    Separate context and system message in instruction formats (#4499) oobabooga 2023-11-07 20:02:58 -03:00
  • 322c170566 Document logits_all oobabooga 2023-11-07 14:45:11 -08:00
  • 5c0559da69 Training: fix .txt files now showing in dropdowns oobabooga 2023-11-07 14:41:11 -08:00
  • af3d25a503 Disable logits_all in llamacpp_HF (makes processing 3x faster) oobabooga 2023-11-07 14:35:48 -08:00
  • 5c3eb22ce6 Bump llama-cpp-python to 0.2.14 oobabooga 2023-11-07 14:20:17 -08:00
  • 3fc505dc0f Document unused parameters oobabooga 2023-11-07 08:56:09 -08:00
  • 3d59346871 Implement echo/suffix parameters oobabooga 2023-11-07 08:43:45 -08:00
  • cee099f131 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev oobabooga 2023-11-07 08:25:22 -08:00
  • 48c9c31440 Document the "preset" option in the API oobabooga 2023-11-07 08:23:17 -08:00
  • d59f1ad89a
    Update README.md oobabooga 2023-11-07 13:05:06 -03:00
  • 0c440877de
    Update 12 - OpenAI API.md oobabooga 2023-11-07 12:59:40 -03:00
  • 55dc9845cb
    Update 12 - OpenAI API.md oobabooga 2023-11-07 12:51:41 -03:00
  • b0b999dd68 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev oobabooga 2023-11-07 07:46:08 -08:00
  • 2bda1a9c9b Mention --api-key oobabooga 2023-11-07 07:45:55 -08:00
  • cc04abda49
    Update 12 - OpenAI API.md oobabooga 2023-11-07 12:40:52 -03:00
  • ddca6948b2
    Update 12 - OpenAI API.md oobabooga 2023-11-07 12:39:59 -03:00
  • 40e73aafce
    Update 12 - OpenAI API.md oobabooga 2023-11-07 12:38:39 -03:00
  • 6ec997f195
    Update 12 - OpenAI API.md oobabooga 2023-11-07 12:36:52 -03:00
  • 15d4ea180d Merge remote-tracking branch 'refs/remotes/origin/dev' into dev oobabooga 2023-11-07 07:35:36 -08:00
  • b2afdda4e8 Add more API examples oobabooga 2023-11-07 07:35:04 -08:00
  • 349604458b
    Update 12 - OpenAI API.md (#4501) Morgan Cheng 2023-11-07 22:22:17 +08:00
  • fd893baba1
    Bump optimum from 1.13.1 to 1.14.0 (#4492) dependabot[bot] 2023-11-07 00:13:41 -03:00
  • 18739c8b3a
    Update peft requirement from ==0.5.* to ==0.6.* (#4494) dependabot[bot] 2023-11-07 00:12:59 -03:00
  • 79b3f5a546
    Add /v1/internal/stop-generation to OpenAI API (#4498) oobabooga 2023-11-07 00:10:42 -03:00
  • 97c21e5667 Don't strip leading spaces in OpenAI API oobabooga 2023-11-06 19:09:41 -08:00
  • 4a45dc4041 Reorder the parameters in the FastAPI documentation oobabooga 2023-11-06 09:55:36 -08:00
  • 1fba6db69f
    Merge pull request #4488 from oobabooga/dev oobabooga 2023-11-06 12:18:55 -03:00
  • 0ed6a17ed4 Update warning oobabooga 2023-11-06 07:17:49 -08:00
  • 0db81355bc Reorder a parameter oobabooga 2023-11-06 07:11:49 -08:00
  • b87c6213ae Remove obsolete endpoint oobabooga 2023-11-06 05:45:45 -08:00
  • fcc9114b58 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev oobabooga 2023-11-06 05:38:47 -08:00
  • ceb8c92dfc
    Update 12 - OpenAI API.md oobabooga 2023-11-06 10:38:22 -03:00
  • 28fd535f9c Make chat API more robust oobabooga 2023-11-06 05:22:01 -08:00
  • 5b5ef57049 Remove file oobabooga 2023-11-05 21:39:59 -08:00
  • ec17a5d2b7
    Make OpenAI API the default API (#4430) oobabooga 2023-11-06 02:38:29 -03:00
  • 84d957ba62
    [Fix] fix openai embedding_model loading as str (#4147) 俞航 2023-11-06 07:42:45 +08:00
  • e18a0460d4
    fix openai extension not working because of absent new defaults (#4477) kabachuha 2023-11-04 22:12:51 +03:00
  • b7a409ef57
    Merge pull request #4476 from oobabooga/dev oobabooga 2023-11-04 15:04:43 -03:00
  • fb3bd0203d Update docs oobabooga 2023-11-04 11:02:24 -07:00
  • 1d8c7c1fc4 Update docs oobabooga 2023-11-04 11:01:15 -07:00
  • b5c53041b8
    Merge pull request #4475 from oobabooga/dev oobabooga 2023-11-04 14:19:55 -03:00
  • 40f7f37009 Update requirements oobabooga 2023-11-04 10:12:06 -07:00
  • 2081f43ac2
    Bump transformers to 4.35.* (#4474) Orang 2023-11-05 00:00:24 +07:00
  • 4766a57352
    transformers: add use_flash_attention_2 option (#4373) feng lui 2023-11-05 00:59:33 +08:00
  • add359379e
    fixed two links in the ui (#4452) wouter van der plas 2023-11-04 17:41:42 +01:00
  • cfbd108826
    Bump AWQ to 0.1.6 (#4470) Casper 2023-11-04 17:09:41 +01:00
  • aa5d671579
    Add temperature_last parameter (#4472) oobabooga 2023-11-04 13:09:07 -03:00
  • 1ab8700d94 Change frequency/presence penalty ranges oobabooga 2023-11-03 17:38:19 -07:00
  • 45fcb60e7a Make truncation_length_max apply to max_seq_len/n_ctx oobabooga 2023-11-03 11:29:31 -07:00
  • 7f9c1cbb30 Change min_p default to 0.0 oobabooga 2023-11-03 08:25:22 -07:00
  • 4537853e2c Change min_p default to 1.0 oobabooga 2023-11-03 08:13:50 -07:00
  • 367e5e6e43
    Implement Min P as a sampler option in HF loaders (#4449) kalomaze 2023-11-02 14:32:51 -05:00
  • fcb7017b7a Remove a checkbox oobabooga 2023-11-02 12:24:09 -07:00
  • fdcaa955e3
    transformers: Add a flag to force load from safetensors (#4450) Julien Chaumond 2023-11-02 20:20:54 +01:00
  • c0655475ae Add cache_8bit option oobabooga 2023-11-02 11:23:04 -07:00