github vercel/ai ai@3.1.0-canary.2

latest releases: @ai-sdk/anthropic@0.0.14, ai@3.1.12, ai@3.1.11...
pre-release2 months ago

ai

3.0.12

Patch Changes

  • b99b008: fix(ai/rsc): avoid appending boundary if the same reference was passed

3.0.11

Patch Changes

  • ce009e2: Added OpenAI assistants streaming.
  • 3f9bf3e: Updates types to OpenAI SDK 4.29.0

3.0.10

Patch Changes

  • 33d261a: fix(ai/rsc): Fix .append() behavior

3.0.9

Patch Changes

  • 81ca3d6: fix(ai/rsc): improve .done() argument type

3.0.8

Patch Changes

  • a94aab2: ai/rsc: optimize streamable value stream size

3.0.7

Patch Changes

  • 9a9ae73: feat(ai/rsc): readStreamableValue

3.0.6

Patch Changes

  • 1355ad0: Fix: experimental_onToolCall is called with parsed tool args
  • 9348f06: ai/rsc: improve dev error and warnings by trying to detect hanging streams
  • 8be9404: fix type resolution

3.0.5

Patch Changes

3.0.4

Patch Changes

  • 7962862: fix useActions type inference
  • aab5324: Revert "fix(render): parse the args based on the zod schema"
  • fe55612: Bump OpenAI dependency to 4.28.4; fix type error in render

3.0.3

Patch Changes

  • 4d816ca: fix(render): parse the args based on the zod schema
  • d158a47: fix potential race conditions

3.0.2

Patch Changes

  • 73bd06e: fix(useActions): return typed object

3.0.1

Patch Changes

  • ac20a25: ai/rsc: fix text response and async generator
  • b88778f: Added onText callback for text tokens.

3.0.0

Major Changes

2.2.37

Patch Changes

  • a6b5764: Add support for Mistral's JavaScript SDK

2.2.36

Patch Changes

  • 141f0ce: Fix: onFinal callback is invoked with text from onToolCall when onToolCall returns string

2.2.35

Patch Changes

  • b717dad: Adding Inkeep as a stream provider

2.2.34

Patch Changes

  • 2c8ffdb: cohere-stream: support AsyncIterable
  • ed1e278: Message annotations handling for all Message types

2.2.33

Patch Changes

  • 8542ae7: react/use-assistant: add onError handler
  • 97039ff: OpenAIStream: Add support for the Azure OpenAI client library

2.2.32

Patch Changes

  • 7851fa0: StreamData: add annotations and appendMessageAnnotation support

2.2.31

Patch Changes

  • 9b89c4d: react/use-assistant: Expose setInput
  • 75751c9: ai/react: Add experimental_onToolCall to useChat.

2.2.30

Patch Changes

  • ac503e0: ai/solid: add chat request options to useChat
  • b78a73e: Add GoogleGenerativeAIStream for Gemini support
  • 5220336: ai/svelte: Add experimental_onToolCall to useChat.
  • ef99062: Add support for the Anthropic message API
  • 5220336: Add experimental_onToolCall to OpenAIStream.
  • ac503e0: ai/vue: add chat request options to useChat

2.2.29

Patch Changes

  • 5a9ae2e: ai/prompt: add experimental_buildOpenAIMessages to validate and cast AI SDK messages to OpenAI messages

2.2.28

Patch Changes

  • 07a679c: Add data message support to useAssistant & assistantResponse.
  • fbae595: ai/react: api functions are no longer used as a cache key in useChat

2.2.27

Patch Changes

  • 0fd1205: ai/vue: Add complex response parsing and StreamData support to useCompletion
  • a7dc746: experimental_useAssistant: Expose extra fetch options
  • 3dcf01e: ai/react Add data support to useCompletion
  • 0c3b338: ai/svelte: Add complex response parsing and StreamData support to useCompletion
  • 8284777: ai/solid: Add complex response parsing and StreamData support to useCompletion

2.2.26

Patch Changes

  • df1ad33: ai/vue: Add complex response parsing and StreamData support to useChat
  • 3ff8a56: Add generateId to use-chat params to allow overriding message ID generation
  • 6c2a49c: ai/react experimental_useAssistant() submit can be called without an event
  • 8b4f7d1: ai/react: Add complex response parsing and StreamData support to useCompletion

2.2.25

Patch Changes

  • 1e61c69: chore: specify the minimum react version to 18
  • 6aec2d2: Expose threadId in useAssistant
  • c2369df: Add AWS Bedrock support
  • 223fde3: ai/svelte: Add complex response parsing and StreamData support to useChat

2.2.24

Patch Changes

  • 69ca8f5: ai/react: add experimental_useAssistant hook and experimental_AssistantResponse
  • 3e2299e: experimental_StreamData/StreamingReactResponse: optimize parsing, improve types
  • 70bd2ac: ai/solid: add experimental_StreamData support to useChat

2.2.23

Patch Changes

  • 5a04321: add StreamData support to StreamingReactResponse, add client-side data API to react/use-chat

2.2.22

Patch Changes

  • 4529831: ai/react: Do not store initialMessages in useState
  • db5378c: experimental_StreamData: fix data type to be JSONValue

2.2.21

Patch Changes

  • 2c8d4bd: Support openai@4.16.0 and later

2.2.20

Patch Changes

  • 424d5ee: experimental_StreamData: fix trailing newline parsing bug in decoder
  • c364c6a: cohere: fix closing cohere stream, avoids response from hanging

2.2.19

Patch Changes

  • 699552d: add experimental_StreamingReactResponse

2.2.18

Patch Changes

  • 0bd27f6: react/use-chat: allow client-side handling of function call without following response

2.2.17

Patch Changes

  • 5ed581d: Use interface instead of type for Message to allow declaration merging
  • 9adec1e: vue and solid: fix including function_call and name fields in subsequent requests

2.2.16

Patch Changes

2.2.15

Patch Changes

  • c5d1857: fix: return complete response in onFinish when onCompletion isn't passed
  • c5d1857: replicate-stream: fix types for replicate@0.20.0+

2.2.14

Patch Changes

  • 6229d6b: openai: fix OpenAIStream types with openai@4.11+

2.2.13

Patch Changes

  • a4a997f: all providers: reset error message on (re)submission

2.2.12

Patch Changes

  • cb181b4: ai/vue: wrap body with unref to support reactivity

2.2.11

Patch Changes

  • 2470658: ai/react: fix: handle partial chunks in react getStreamedResponse when using experimental_StreamData

2.2.10

Patch Changes

  • 8a2cbaf: vue/use-completion: fix: don't send network request for loading state"
  • bbf4403: langchain-stream: return langchain writer from LangChainStream

2.2.9

Patch Changes

  • 3fc2b32: ai/vue: fix: make body parameter reactive

2.2.8

Patch Changes

  • 26bf998: ai/react: make reload/complete/append functions stable via useCallback

2.2.7

Patch Changes

  • 2f97630: react/use-chat: fix aborting clientside function calls too early
  • 1157340: fix: infinite loop for experimental stream data (#484)

2.2.6

Patch Changes

  • e5bf68d: react/use-chat: fix experimental functions returning proper function messages

    Closes #478

2.2.5

Patch Changes

  • e5bf68d: react/use-chat: fix experimental functions returning proper function messages

    Closes #478

2.2.4

Patch Changes

  • 7b389a7: fix: improve safety for type check in openai-stream

2.2.3

Patch Changes

  • 867a3f9: Fix client-side function calling (#467, #469)

    add Completion type from the openai SDK to openai-stream (#472)

2.2.2

Patch Changes

  • 84e0cc8: Add experimental_StreamData and new opt-in wire protocol to enable streaming additional data. See #425.

    Changes onCompletion back to run every completion, including recursive function calls. Adds an onFinish callback that runs once everything has streamed.

    If you're using experimental function handlers on the server and caching via onCompletion,
    you may want to adjust your caching code to account for recursive calls so the same key isn't used.

    let depth = 0
    
    const stream = OpenAIStream(response, {
        async onCompletion(completion) {
          depth++
          await kv.set(key + '_' + depth, completion)
          await kv.expire(key + '_' + depth, 60 * 60)
        }
      })
    

2.2.1

Patch Changes

  • 04084a8: openai-stream: fix experimental_onFunctionCall types for OpenAI SDK v4

2.2.0

Minor Changes

  • dca1ed9: Update packages and examples to use OpenAI SDK v4

2.1.34

Patch Changes

  • c2917d3: Add support for the Anthropic SDK, newer Anthropic API versions, and improve Anthropic error handling

2.1.33

Patch Changes

  • 4ef8015: Prevent isLoading in vue integration from triggering extraneous network requests

2.1.32

Patch Changes

  • 5f91427: ai/svelte: fix isLoading return value

2.1.31

Patch Changes

2.1.30

Patch Changes

  • 4df2a49: Fix termination of ReplicateStream by removing the terminating {}from output

2.1.29

Patch Changes

  • 3929a41: Add ReplicateStream helper

2.1.28

Patch Changes

  • 9012e17: react/svelte/vue: fix making unnecessary SWR request to API endpoint

2.1.27

Patch Changes

  • 3d29799: React/Svelte/Vue: keep isLoading in sync between hooks with the same ID.

    React: don't throw error when submitting

2.1.26

Patch Changes

  • f50d9ef: Add experimental_buildLlama2Prompt helper for Hugging Face

2.1.25

Patch Changes

  • 877c16f: ai/react: don't throw error if onError is passed

2.1.24

Patch Changes

  • f3f5866: Adds SolidJS support and SolidStart example

2.1.23

Patch Changes

  • 0ebc2f0: streams/openai-stream: don't call onStart/onCompletion when recursing

2.1.22

Patch Changes

  • 9320e95: Add (experimental) prompt construction helpers for StarChat and OpenAssistant
  • e3a7ec8: Support <|end|> token for StarChat beta in huggingface-stream

2.1.21

Patch Changes

  • 561a49a: Providing a function to function_call request parameter of the OpenAI Chat Completions API no longer breaks OpenAI function stream parsing.

2.1.20

Patch Changes

  • e361114: OpenAI functions: allow returning string in callback

2.1.19

Patch Changes

  • e4281ca: Add experimental server-side OpenAI function handling

2.1.18

Patch Changes

  • 6648b21: Add experimental client side OpenAI function calling to Svelte bindings
  • e5b983f: feat(streams): add http error handling for openai

2.1.17

Patch Changes

  • 3ed65bf: Remove dependency on node crypto API

2.1.16

Patch Changes

  • 8bfb43d: Fix svelte peer dependency version

2.1.15

Patch Changes

  • 4a2b978: Update cohere stream and add docs

2.1.14

Patch Changes

  • 3164adb: Fix regression with generated ids

2.1.13

Patch Changes

  • fd82961: Use rfc4122 IDs when generating chat/completion IDs

2.1.12

Patch Changes

2.1.11

Patch Changes

  • 8bf637a: Fix langchain handlers so that they now are correctly invoked and update examples and docs to show correct usage (passing the handlers to llm.call and not the model itself).

2.1.10

Patch Changes

  • a7b3d0e: Experimental support for OpenAI function calling

2.1.9

Patch Changes

  • 9cdf968: core/react: add Tokens react server component

2.1.8

Patch Changes

  • 44d9879: Support extra request options in chat and completion hooks

2.1.7

Patch Changes

  • bde3898: Allow an async onResponse callback in useChat/useCompletion

2.1.6

Patch Changes

  • 23f0899: Set stream: true when decoding streamed chunks

2.1.5

Patch Changes

  • 89938b0: Provider direct callback handlers in LangChain now that CallbackManager is deprecated.

2.1.4

Patch Changes

  • c16d650: Improve type saftey for AIStream. Added JSDoc comments.

2.1.3

Patch Changes

  • a9591fe: Add createdAt on user input message in useChat (it was already present in assistant messages)

2.1.2

Patch Changes

2.1.1

Patch Changes

  • 9fdb51a: fix: add better typing for store within svelte implementation (#104)

2.1.0

Minor Changes

  • 71f9c51: This adds Vue support for ai via the ai/vue subpath export. Vue composables useChat and useCompletion are provided.

Patch Changes

2.0.1

Patch Changes

  • be90740: - Switches LangChainStream helper callback handler to return use handleChainEnd instead of handleLLMEnd so as to work with sequential chains

2.0.0

Major Changes

0.0.14

Patch Changes

  • c6586a2: Add onError callback, include response text in error if response is not okay

0.0.13

Patch Changes

  • c1f4a91: Throw error when provided AI response isn't valid

0.0.12

Patch Changes

0.0.11

Patch Changes

  • a6bc35c: fix package exports for react and svelte subpackages

0.0.10

Patch Changes

0.0.9

Patch Changes

  • 78477d3: - Create /react sub-package.
    • Create import { useChat, useCompletion } from 'ai/react' and mark React as an optional peer dependency so we can add more framework support in the future.
    • Also renamed set to setMessages and setCompletion to unify the API naming as we have setInput too.
    • Added an sendExtraMessageFields field to useChat that defaults to false, to prevent OpenAI errors when id is not filtered out.
  • c4c1be3: useCompletion.handleSubmit does not clear the input anymore
  • 7de2185: create /react export

0.0.8

Patch Changes

  • fc83e95: Implement new start-of-stream newline trimming
  • 2c6fa04: Optimize callbacks TransformStream to be more memory efficient when onCompletion is not specified

0.0.7

Patch Changes

  • fdfef52: - Splits the EventSource parser into a reusable helper
    • Uses a TransformStream for this, so the stream respects back-pressure
    • Splits the "forking" stream for callbacks into a reusable helper
    • Changes the signature for customParser to avoid Stringify -> Encode -> Decode -> Parse round trip
    • Uses ?.() optional call syntax for callbacks
    • Uses string.includes to perform newline checking
    • Handles the null res.body case
    • Fixes Anthropic's streaming responses

0.0.6

Patch Changes

  • d70a9e7: Add streamToResponse
  • 47b85b2: Improve abortController and callbacks of useChat
  • 6f7b43a: Export UseCompletionHelpers as a TypeScript type alias

0.0.5

Patch Changes

  • 4405a8a: fix duplicated 'use client' directives

0.0.4

Patch Changes

  • b869104: Added LangChainStream, useCompletion, and useChat

0.0.3

Patch Changes

0.0.2

Patch Changes

0.0.1

Patch Changes

0.0.2

Patch Changes

Don't miss a new ai release

NewReleases is sending notifications on new releases.