0.1.5 • Published 5 months ago

@ai-d/aid v0.1.5

Weekly downloads
-
License
MIT
Repository
github
Last release
5 months ago

Aid: TypeScript Library for Typed LLM Interactions

A.I. :D

Aid is a TypeScript library designed for developers working with Large Language Models (LLMs) such as OpenAI's GPT-4 (including Vision) and GPT-3.5. The library focuses on ensuring consistent, typed outputs from LLM queries, enhancing the reliability and usability of LLM responses. Advanced users can leverage few-shot examples for more sophisticated use cases. It provides a structured and type-safe way to interact with LLMs.

Features

  • Typed Response: Aid leverages TypeScript and JSON Schema to ensure consistent, reliable outputs from LLMs, adheres to the predefined schema.
  • Task Based: Easily define custom tasks with specific input and output types, streamlining the process of LLM interactions.
  • Few-Shot Learning Support: Allows for the provision of few-shot prompt examples to guide the LLM in producing the desired output.
  • Visual Task Support: Includes support for visual tasks with image inputs, harnessing the power of OpenAI's GPT-4 Vision. Example
  • OpenAI Integration: Integrates with OpenAI's official library to provide a seamless experience.
  • Customizable: Allows for customization LLM models, just implement the QueryEngine function. Example

Installation

pnpm install @ai-d/aid

Usage

Basic Setup

First, import the necessary modules and set up your OpenAI instance:

import { OpenAI } from "openai";
import { Aid } from "@ai-d/aid";

const openai = new OpenAI({ apiKey: "your-api-key" });
const aid = Aid.from(openai, { model: "gpt-4-1106-preview" });
import { OpenAI } from "openai";
import { Aid, OpenAIQuery } from "@ai-d/aid";

const openai = new OpenAI({ apiKey: "your-api-key" });
const aid = Aid.vision(
    OpenAIQuery(openai, { model: "gpt-4-vision-preview", max_tokens: 2048 }),
);

For example, Cohere's Command.

import { Aid, CohereQuery } from "@ai-d/aid";

const aid = Aid.chat(
    CohereQuery(COHERE_TOKEN, { model: "command" }),
);

You can implement your own QueryEngine function.

Creating a Custom Task

Define a custom task with expected output types:

import { z } from "zod";

const analyze = aid.task(
    "Summarize and extract keywords",
    z.object({
        summary: z.string().max(300),
        keywords: z.array(z.string().max(30)).max(10),
    }),
);
const analyze = aid.task(
    "Analyze the person in the image",
    z.object({
        gender: z.enum(["boy", "girl", "other"]),
        age: z.enum(["child", "teen", "adult", "elderly"]),
        emotion: z.enum(["happy", "sad", "angry", "surprised", "neutral"]),
        clothing: z.string().max(100),
        background: z.string().max(100),
    }),
);

Executing a Task

Execute the task and handle the output:

const { result } = await analyze("Your input here, e.g. a news article");
console.log(result); // { summary: "...", keywords: ["...", "..."] }
const datauri = `data:image/png;base64,${fs.readFileSync("path/to/image.png" "base64")}`;

const { result } = await analyze({ images: [{ url: datauri }] });
console.log(result); // { "gender": "boy", "age": "teen", ... }

Advanced Usage with Few-Shot Examples

For more complex scenarios, you can use few-shot examples:

const run_advanced_task = aid.task(
    "Some Advanced Task",
    z.object({
        // Define your output schema here
    }),
    {
        examples: [
            // Provide few-shot examples here
        ],
    }
);

Formulation

Case Parameter -> (join) Task Defination -> (join) Format Constraint -> (perform) Query

Query and Format Constraint are defined and implemented by the QueryEngine and FormatEngine.

Task Defination is defined by the user with task method. Task Goal, Expected Schema, Examples, etc.

Case Parameter is defined by the user on each single call. Text, Image, etc.

Contributing

Contributions are welcome! Please submit pull requests with any bug fixes or feature enhancements.