88export default async function server(request) {
89 if (request.method === "POST" && new URL(request.url).pathname === "/chat") {
90 const { OpenAI } = await import("https://esm.town/v/std/openai");
91 const openai = new OpenAI();
92
93 const { messages } = await request.json();
94
95 try {
96 const completion = await openai.chat.completions.create({
97 messages,
98 model: "gpt-4o-mini",
1import { OpenAI } from "https://esm.town/v/std/openai";
2export default async function(req: Request): Promise<Response> {
3 const openai = new OpenAI();
4 const stream = await openai.chat.completions.create({
5 stream: true,
6 messages: [{ role: "user", content: "Write a poem in the style of beowulf about the DMV" }],
231 const { query } = await request.json();
232
233 const { OpenAI } = await import("https://esm.town/v/std/openai");
234 const openai = new OpenAI();
235
236 const controller = new AbortController();
249 async start(controller) {
250 try {
251 const completion = await openai.chat.completions.create({
252 messages: [
253 {
353export default async function server(request: Request): Promise<Response> {
354 try {
355 const { OpenAI } = await import("https://esm.town/v/std/openai");
356
357 if (request.method === "POST" && new URL(request.url).pathname === "/chat") {
358 const openai = new OpenAI();
359 const { messages } = await request.json();
360
407 }
408
409 // OpenAI query
410 try {
411 const completion = await openai.chat.completions.create({
412 model: "gpt-4o-mini",
413 messages: messages,
423 });
424 } catch (error) {
425 console.error("OpenAI API Error:", error);
426 return new Response("Sorry, something went wrong.", {
427 status: 500,
72export default async function server(request: Request): Promise<Response> {
73 if (request.method === "POST") {
74 const { OpenAI } = await import("https://esm.town/v/std/openai");
75 const openai = new OpenAI();
76
77 try {
78 const { messages } = await request.json();
79 const completion = await openai.chat.completions.create({
80 messages: messages,
81 model: "gpt-4o-mini",
335 "value": {
336 "stringValue":
337 "{\"OPENAI_API_KEY\":\"sk-proj-uoRQ71Oaz16IbDJRJXY5T3BlbkFJgqSSf3Xglg0V0V2FhFrO\",\"ANTHROPIC_API_KEY\":\"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\",\"DATABASE_URL\":\"postgresql://neondb_owner:XaenL6g5WApI@ep-proud-grass-a5r8g74c.us-east-2.aws.neon.tech/neondb?sslmode=require\",\"FPX_ENDPOINT\":\"http://localhost:8788/v1/traces\",\"FPX_LOG_LEVEL\":\"debug\",\"GOOSE_AVATARS\":{}}",
338 },
339 },
785 "key": "fpx.http.request.env",
786 "value": {
787 "stringValue": "{\"OPENAI_API_KEY\":\"sk-proj-uoRQ71Oaz16IbDJRJXY5T3BlbkFJgqSSf3Xglg0V0V2FhFrO\",\"ANTHROPIC_API_KEY\":\"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\",\"DATABASE_URL\":\"postgresql://neondb_owner:XaenL6g5WApI@ep-proud-grass-a5r8g74c.us-east-2.aws.neon.tech/neondb?sslmode=require\",\"FPX_ENDPOINT\":\"http://localhost:8788/v1/traces\",\"FPX_LOG_LEVEL\":\"debug\",\"GOOSE_AVATARS\":{}}"
788 }
789 },
108export default async function server(request: Request): Promise<Response> {
109 if (request.method === "POST" && new URL(request.url).pathname === "/chat") {
110 const { OpenAI } = await import("https://esm.town/v/std/openai");
111 const openai = new OpenAI();
112
113 const { messages } = await request.json();
114
115 try {
116 const completion = await openai.chat.completions.create({
117 messages: messages,
118 model: "gpt-4o-mini",
108export default async function server(request: Request): Promise<Response> {
109 if (request.method === "POST" && new URL(request.url).pathname === "/chat") {
110 const { OpenAI } = await import("https://esm.town/v/std/openai");
111 const openai = new OpenAI();
112
113 const { messages } = await request.json();
114
115 try {
116 const completion = await openai.chat.completions.create({
117 messages: messages,
118 model: "gpt-4o-mini",
108export default async function server(request: Request): Promise<Response> {
109 if (request.method === "POST" && new URL(request.url).pathname === "/chat") {
110 const { OpenAI } = await import("https://esm.town/v/std/openai");
111 const openai = new OpenAI();
112
113 const { messages } = await request.json();
114
115 try {
116 const completion = await openai.chat.completions.create({
117 messages: messages,
118 model: "gpt-4o-mini",
108export default async function server(request: Request): Promise<Response> {
109 if (request.method === "POST" && new URL(request.url).pathname === "/chat") {
110 const { OpenAI } = await import("https://esm.town/v/std/openai");
111 const openai = new OpenAI();
112
113 const { messages } = await request.json();
114
115 try {
116 const completion = await openai.chat.completions.create({
117 messages: messages,
118 model: "gpt-4o-mini",