aimyapi
v0.2.1
Published
AI My API! Use LLMs to create small programs that safely* run and make calls against your API. Similar to ChatGPT's Code Interpreter.
Downloads
5
Readme
AI My API
Build AI assistants that use your API's and respond to natural language queries by writing small programs that are executed in a QuickJS sandbox. Similar to chatGPT Code Interpreter but running on your servers and calling your own API.
AIMyAPI writes the glue between your various API functions unlocking new functionality without having to write any code. With the right building blocks, AIMyAPI can even invent new functionality you didn't originally plan for - for example, being able to sort or filter your data in new ways, email or text yourself results, or combine the results of different operations and API's.
Uses the OpenAI API and requires an API key.
Quickstart
Install the package
npm i aimyapi --save
Important: create a .env file in the root of your project and add your OpenAI api key:
OPENAI_API_KEY=XXXXXXXX
Simply define your strongly-typed typescript API:
sales_api.ts
// Sales data for each month, per business unit
export interface SalesData {
sales: number;
month: number; // 1...12
year: number;
businessUnit: "Advertising" | "Hardware" | "Software";
}
export interface APIInterface {
getSalesData(): SalesData[];
// Returns an array of sales data.
print(text: string): void;
// Prints the specified text to the user.
};
// Global object you will use for accessing the APIInterface
declare global {
var api: APIInterface;
}
Implement it (sales_api_impl.ts
):
export class API implements APIInterface {
getSalesData() {
...
Document it (sales_api.md
):
Examples of how you should respond:
user: How are you today?
assistant:
api.print("I'm great, thanks for asking!");
Instantiate your api and AIMyAPI. AIMyAPI will wrap your API so that it can be called from within the QuickJS sandbox.
import { AIMyAPI } from "aimyapi"
const api = new API();
const aimyapi = await AIMyAPI.createWithAPI({
apiObject: api,
apiExports: APIExports,
apiDefFilePath: path.join(__dirname, "./sales_api.ts"),
apiDocsPath: path.join(__dirname, "./sales_api.md"),
})
And now you can make requests against the API in natural language. Any code generated by the LLM will be run within the sandbox. The LLM will ONLY have access to functions provided within your API and cannot use javascript to make http requests etc.
await aimyapi.processRequest("Which business unit had the highest sales in 2020?");
This would produce the following code generated by the LLM prompt, which is run in the sandbox and makes calls on your API object:
import * as ApiDefs from './api.ts'
(async() => {
try {
let salesData = api.getSalesData();
let maxSales = 0;
let businessUnitWithMaxSales = "";
for (let i = 0; i < salesData.length; i++) {
if (salesData[i].year === 2020 && salesData[i].sales > maxSales) {
maxSales = salesData[i].sales;
businessUnitWithMaxSales = salesData[i].businessUnit;
}
}
api.print(\`The business unit with the highest sales in 2020 was ${businessUnitWithMaxSales}.\`);
} catch(e) {
console.error(e);
}
})();
This example does not take advantage of the chat history and treats every request as a fresh chat. See example2
for a more complex example that uses the chat history.
APIs should be kept simple and understandable. If you have a complicated API you may want to write a simpler facade for it.
Advantages and disadvantages of generating code
Currently there is a lot of hype around "agents" - while agents are extremely cool they have a couple drawbacks vs. the code generation approach that AIMyAPI uses to do some of the same things.
Advantages of AIMyAPI over the Agent (haystack) approach:
- Open AI's LLMs today are already really good at writing code, but not necessarily trained on the 'haystack' approach.
- Data from your API can be kept private as long as you don't add it to the chat history or the prompt. The LLM is writing the operations to manipulate the data but doesn't need to see the data itself to do that.
- Writing code means that AIMyAPI can do math and algorithms out of the box - something LLMs by themselves struggle at. For instance, asking it to do a simple calculation will result in it writing code to do the calculation. Agents can do this using 'tools', but that means you need to do the work to implement and explain these tools wheras they come built-in in a programming language.
- Agents typically only do one thing at a time and require several OpenAI API calls to do a single loop. AIMyAPI can write an entire program that does several things in one shot.
Disadvantages:
- This is not an iterative approach: AIMyAPI must have a full understanding of the problem and the API in order write code against it. Haystack may be better for problems with unknowns.
- The LLM can misunderstand your intentions, the api or write incorrect code or reference things that don't exist in the sandbox.
- It is not currently set up to be able to troubleshoot or fix it's own code (this may be possible though)
- Limits on the context for OpenAI's LLMs mean you can't have a super big API or lots of docs/examples (maybe GPT4's larger context will make this a non-issue - I don't have access yet).
- Due to it's training, gpt3.5-turbo tends to want to reply with commentary sometimes, not just code. But examples mostly fix this issue. The
davinci-003
model with the completion api can be prompted to just reply with code, but it's 10x more expensive.
Installation
npm install
Important: create a .env file in the root of your project add your open ai key:
OPENAI_API_KEY=XXXXXXXX
API Reference for AIMyAPI
createWithAPI(options: AIMyAPIOptions): Promise
Primary function you will use to create a module for working your API API. For a simple example with no chat history, see example1. For an example with chat history see example2.
options
apiObject
: An object representing the API to be used.apiExports
: An object representing the exported functions and variables from the API.apiGlobals
(optional): An object representing global variables for the API.apiDefFilePath
: A string representing the full path to the API definition file.apiGlobalName
(optional): A string representing the name of the global variable for the API.apiDocsPath
(optional): A string representing the path to the documentation for the API.debug
(optional): A boolean indicating whether to output debugging information.
The AIMyAPI module (type: CreateWithAPIExports)
The module created by the createWithAPI
function:
Methods
generateCode(queryText: string, userChatHistory: ChatCompletionRequestMessage[]): Promise<string>
: A function that generates code to be run in the sandbox for the given query text and user chat history.runCode(task: string): Promise<void>
: A function that runs the generated code in the sandbox.processRequest(userQuery: string, context?: object): Promise<string>
: A single function that generates and runs code assuming no chat history.
Examples
Download or clone the repo. From the root folder run the following.
Example 1 - sales database
A simple api example where the LLM allows you to ask questions and compute things about a dataset
npm run example1
import { AIMyAPI } from "../lib";
import { API } from "./sales_api_impl";
import * as APIExports from "./sales_api";
import path from "path";
// Simpler example using no support for chat history;
(async () => {
const api = new API();
const aimyapi = await AIMyAPI.createWithAPI({
apiObject: api,
apiGlobalName: "api", // Should match whatever you declared as your global in your api.
apiExports: APIExports,
apiDefFilePath: path.join(__dirname, "./sales_api.ts"),
apiDocsPath: path.join(__dirname, "./sales_api.md"),
debug: false,
})
console.log("Print 'hello' three times.")
await aimyapi.processRequest("Print 'hello' three times.");
console.log("What was our best month in 2020")
await aimyapi.processRequest("What was our best month in 2020");
console.log("Which business unit had the highest sales in 2020?")
await aimyapi.processRequest("Which business unit had the highest sales in 2020?");
console.log("Compute the total sales for each month in 2020 then email the top 3 to [email protected]")
await aimyapi.processRequest("Compute the total sales for each month in 2020 then email the top 3 to [email protected]");
})();
Output:
Print 'hello' three times.
Generate Task: 4.086s
hello
hello
hello
Run Code: 379.328ms
What was our best month in 2020
Generate Task: 11.345s
Our best month in 2020 was month 8 with 4400 sales.
Run Code: 119.145ms
Which business unit had the highest sales in 2020?
Generate Task: 11.477s
The business unit with the highest sales in 2020 was Hardware
Run Code: 124.522ms
Compute the total sales for each month in 2020 then email the top 3 to [email protected]
Generate Task: 15.648s
Sending email to "[email protected]" with subject "Top 3 Sales Months in 2020" and body
"Top 3 sales months in 2020:
Month 8: $4400
Month 7: $4300
Month 6: $4000"
Run Code: 1.200s
Example 2 - Food ordering bot
A more complex food ordering example which incorporates chat history
npm run example2
Code:
import { AIMyAPI } from "../lib";
import { OrderingAPI } from "./ordering_api_impl";
import * as APIExports from "./ordering_api";
import path from "path";
// More complex fast food ordering example that uses chat history.
// Fast food ordering example.
(async () => {
const api = new OrderingAPI();
const aimyapi = await AIMyAPI.createWithAPI({
apiObject: api,
apiGlobalName: "orderingApi", // Should match whatever you declared as your global in your ordering api.
apiExports: APIExports,
apiDefFilePath: path.join(__dirname, "./ordering_api.ts"),
apiDocsPath: path.join(__dirname, "./ordering_api.md"),
debug: false,
})
async function runQuery(query:string) {
console.log(`Query: ${query}`)
// generate the code for this query
const codeResult = await aimyapi.generateCode(query, api._getHistory());
api._addMessageToHistory({
content: query,
role: "user",
name: "user",
});
api._addMessageToHistory({
content: '```\n' + codeResult.loggableCode + '\n```',
role: "assistant",
name: "assistant",
});
// run the code in the sandbox
await aimyapi.runCode(codeResult.code);
}
await runQuery("How's the burger here?");
await runQuery("I'd like a plain hamburger and a large fries, a pizza with anchovies and a cola. I'd also like a cheese burger with bacon and avocado.");
await runQuery("I changed my mind, instead of anchovies, can I get a large gluten free pizza with pineapple instead.");
await runQuery("email me the total at [email protected] and complete the order");
})();
Output:
Query: How's the burger here?
Our Hamburger is delicious! It comes with cheese, lettuce, tomato, and onion by default, but you can customize it with any of our toppings.
Query: I'd like a plain hamburger (no cheese) and a large fries, a pizza with anchovies and a cola. I'd also like a cheese burger with bacon and avocado.
added #0 Hamburger... $8.99
lettuce
tomato
onion
Ok, I added a plain hamburger. What else would you like?
added #1 Fries... $2.99
Large... $1.99
Ok, I added large fries. What else would you like?
added #2 Pizza... $10.99
anchovies
Ok, I added a pizza with anchovies. What else would you like?
added #3 Fountain Drink... $1.99
Large... $0.99
Ok, I added a cola. What else would you like?
added #4 Hamburger... $8.99
cheese
lettuce
tomato
onion
cheese
bacon
avocado
Ok, I added a cheese burger with bacon and avocado. Is there anything else I can help you with?
Query: I changed my mind, instead of anchovies, can I get a large gluten free pizza with pineapple instead.
Removed item 2
added #5 Pizza... $10.99
pineapple
Gluten Free Crust... $2.99
Ok, I added a large gluten free pizza with pineapple. Is there anything else I can help you with?
Query: email me the total at [email protected] and complete the order
Sending email to "[email protected]" with subject "Order Total" and body
"Your order total is $39.92."
----
Displaying order:
#0 Hamburger... $8.99
lettuce
tomato
onion
#1 Fries... $2.99
Large... $1.99
#3 Fountain Drink... $1.99
Large... $0.99
#4 Hamburger... $8.99
cheese
lettuce
tomato
onion
cheese
bacon
avocado
#5 Pizza... $10.99
pineapple
Gluten Free Crust... $2.99
-- Total: $39.92
Order completed!
Your order has been completed. Thank you for choosing our restaurant!
Example 3 - interactive food ordering
Interactive ordering example (same api as example2). You can type your own orders on the command line.
npm run example3
Providing examples
Providing examples is super important.
We recommend that you instruct the LLM to return a self executing async function enclosed in ``` with no other commentary. See the example/ordering_api.md
for reference.
It's recommended that you provide at least one example and no other commentary from the assistant, otherwise the LLM's responses will vary pretty widely and probably be unusable.
GPT3.5-turbo also has tendency to respond in words even after being instructed to respond only with code. Examples help alleviate this problem.
Sample example:
Examples of how you should respond:
user: How are you today?
assistant:
\`\`\`
api.print("I'm great, thanks for asking!");
\`\`\`
Aside from the code generation piece, documenting your API with clearly named functions and comments is important so that the LLM understands how it can be used. A big part of using AIMyAPI is testing out usecases and looking at the code that is generated. Problems can be solved by how you name your api, comments in the api, and examples.
A note about history
If you plan to use history, it is vital that you include the generated code in the history. Otherwise, the LLM will start to think that it doesn't need to generate code. Logging is also necessary for keeping track of Id's from past requests so that future requests can leverage that information.
See example2 for more an example on how to do this.
TODO: Add more details explaining how this is done.
Safety notes:
- If the user is giving sensitive information in a query, that information isn't private. Refer to the OpenAI policies and plan accordingly.
- The LLM will do stupid things with your API. You should idiot proof your functions, make sure you sanity check any inputs, etc.
- Even though we are running the code in sandbox, you should still treat anything coming back from the LLM as untrusted code.
- You should never give the LLM access to anything mission critical through your API such as live production data or services.
- Give as few permissions as is necessary - read only access would be preferred.
- Users can trick the LLM to do things that you may not want or intend. Again, access controls are critical. You shouldn't let users access anything they wouldn't normally be able to access via your API.
- Any 'dangerous' operations should involve some kind of human oversight and approval.
- A way to audit and rollback anything the LLM does would be wise. The LLM can misunderstand requests due to false assumptions or incomplete information (or trickery). If you allow it to modify your data it could mess things up.
- By using this library you acknowledge that you take full responsibility for anything that happens as a result of it's use. We cannot gaurantee what will happen based on your instructions to the LLM or how the LLM interprets them etc.
Thoughts:
Could fine tuning on our API and examples allow us to make the prompt smaller? We can only fine tune on the base models though, so we can't use the cheaper gpt3.5-turbo model. Could a smaller / more efficient LLM focused on coding tasks work for this usecase?