diff --git a/README.md b/README.md index cffa51a04..68836b2c9 100644 --- a/README.md +++ b/README.md @@ -43,7 +43,7 @@ It comes packed with **world-class features** like Beam, and is praised for its ### What makes Big-AGI different: -**Intelligence**: with [Beam & Merge](https://big-agi.com/beam) for multi-model de-hallucination, native search, and bleeding-edge AI models like Opus 4.6, Nano Banana Pro, Kimi K2.5 or GPT 5.2 - +**Intelligence**: with [Beam & Merge](https://big-agi.com/beam) for multi-model de-hallucination, native search, and bleeding-edge AI models like Opus 4.6, Nano Banana Pro, Kimi K2.5 or GPT 5.4 - **Control**: with personas, data ownership, requests inspection, unlimited usage with API keys, and *no vendor lock-in* - and **Speed**: with a local-first, over-powered, zero-latency, madly optimized web app. @@ -74,7 +74,7 @@ Purest AI outputs Flow-state interface
-Higly customizable
+Highly customizable
Best-in-class UX @@ -182,8 +182,11 @@ The new architecture is solid and the speed improvements are real.
-What's New in 1.16.1...1.16.10 · 2024-2025 (patch releases) +What's New in 1.16.1...1.16.13 · (patch releases) +- 1.16.13: Docker fix ([#840](https://github.com/enricoros/big-AGI/issues/840)) +- 1.16.12: Dockerfile update ([#840](https://github.com/enricoros/big-AGI/issues/840)) +- 1.16.11: v1 final release, documentation updates - 1.16.10: OpenRouter models support - 1.16.9: Docker Gemini fix, R1 models support - 1.16.8: OpenAI ChatGPT-4o Latest, o1 models support @@ -245,7 +248,7 @@ The new architecture is solid and the speed improvements are real. - New **[Perplexity](https://www.perplexity.ai/)** and **[Groq](https://groq.com/)** integration (thanks @Penagwin). [#407](https://github.com/enricoros/big-AGI/issues/407), [#427](https://github.com/enricoros/big-AGI/issues/427) - **[LocalAI](https://localai.io/models/)** deep integration, including support for [model galleries](https://github.com/enricoros/big-AGI/issues/411) - **Mistral** Large and Google **Gemini 1.5** support -- Performance optimizations: runs [much faster](https://twitter.com/enricoros/status/1756553038293303434?utm_source=localhost:3000&utm_medium=big-agi), saves lots of power, reduces memory usage +- Performance optimizations: runs [much faster](https://x.com/enricoros/status/1756553038293303434?utm_source=localhost:3000&utm_medium=big-agi), saves lots of power, reduces memory usage - Enhanced UX with auto-sizing charts, refined search and folder functionalities, perfected scaling - And with more UI improvements, documentation, bug fixes (20 tickets), and developer enhancements @@ -326,14 +329,15 @@ Delightful UX with latest models exclusive features like Beam for **multi-model Configure 100s of AI models from 20+ providers: -| **AI models** | _supported vendors_ | -|:--------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| -| Opensource Servers | [LocalAI](https://localai.io/) · [Ollama](https://ollama.com/) | -| Local Servers | [LM Studio](https://lmstudio.ai/) (non-open) | -| Multimodal services | [Anthropic](https://anthropic.com) · [AWS Bedrock](https://aws.amazon.com/bedrock/) · [Azure](https://azure.microsoft.com/en-us/products/ai-services/openai-service) · [Google Gemini](https://ai.google.dev/) · [OpenAI](https://platform.openai.com/docs/overview) | +| **AI models** | _supported vendors_ | +|:--------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| +| Opensource Servers | [LocalAI](https://localai.io/) · [Ollama](https://ollama.com/) | +| Local Servers | [LM Studio](https://lmstudio.ai/) (non-open) | +| Multimodal services | [Anthropic](https://anthropic.com) · [AWS Bedrock](https://aws.amazon.com/bedrock/) · [Azure](https://azure.microsoft.com/en-us/products/ai-services/openai-service) · [Google Gemini](https://ai.google.dev/) · [OpenAI](https://platform.openai.com/docs/overview) | | LLM services | [Alibaba](https://www.alibabacloud.com/en/product/modelstudio) · [DeepSeek](https://deepseek.com) · [Groq](https://wow.groq.com/) · [Mistral](https://mistral.ai/) · [Moonshot](https://www.moonshot.cn/) · [OpenPipe](https://openpipe.ai/) · [OpenRouter](https://openrouter.ai/) · [Perplexity](https://www.perplexity.ai/) · [Together AI](https://www.together.ai/) · [xAI](https://x.ai/) · [Z.ai](https://z.ai/) | -| Image services | OpenAI · Google Gemini | -| Speech services | [ElevenLabs](https://elevenlabs.io) · [Inworld](https://inworld.ai) · [OpenAI TTS](https://platform.openai.com/docs/guides/text-to-speech) · LocalAI · Browser (Web Speech API) | +| OpenAI-compatible | Any OpenAI-compatible endpoint - models, pricing, and capabilities are auto-detected | +| Image services | OpenAI · Google Gemini (Nano Banana) · LocalAI | +| Speech services | [ElevenLabs](https://elevenlabs.io) · [Inworld](https://inworld.ai) · [OpenAI TTS](https://platform.openai.com/docs/guides/text-to-speech) · LocalAI · Browser (Web Speech API) | ### Additional Integrations @@ -389,4 +393,4 @@ When you open an issue, our custom AI triage system (powered by [Claude Code](ht MIT License · [Third-Party Notices](src/modules/3rdparty/THIRD_PARTY_NOTICES.md) -**2023-2026** · Enrico Ros × [Big-AGI](https://big-agi.com) +**2023-2026** · [Enrico Ros](https://www.enricoros.com) × [Token Fabrics](https://www.tokenfabrics.com) diff --git a/docs/2024-AI-APIs-Comparison.md b/docs/2024-AI-APIs-Comparison.md index 0af2ebe97..28d52b5b5 100644 --- a/docs/2024-AI-APIs-Comparison.md +++ b/docs/2024-AI-APIs-Comparison.md @@ -1,3 +1,7 @@ +--- +unlisted: true +--- + # AIX dispatch server - API features comparison This is updated as of 2024-07-09, and includes the latest features and capabilities of the three major AI APIs: Anthropic, Gemini, and OpenAI. diff --git a/docs/README.md b/docs/README.md index 2d3f465da..8c16ab285 100644 --- a/docs/README.md +++ b/docs/README.md @@ -10,6 +10,8 @@ Essential guides: - **[FAQ](help-faq.md)**: Common questions and answers - **[Enabling Microphone](help-feature-microphone.md)**: Configure speech recognition in your browser +- **[Data Ownership](help-data-ownership.md)**: How your data is stored and managed +- **[Live File](help-feature-livefile.md)**: Live file attachment feature ## AI Services @@ -26,15 +28,16 @@ How to set up AI models and features in big-AGI. [Google Gemini](https://aistudio.google.com/app/apikey), [Groq](https://console.groq.com/keys), [Mistral](https://console.mistral.ai/api-keys/), + [Moonshot](https://platform.moonshot.cn/console/api-keys), [OpenAI](https://platform.openai.com/api-keys), [OpenPipe](https://app.openpipe.ai/settings), [Perplexity](https://www.perplexity.ai/settings/api), [TogetherAI](https://api.together.xyz/settings/api-keys), - [xAI](http://x.ai/api), + [xAI](https://x.ai/api), [Z.ai](https://z.ai/) - **[Azure OpenAI](config-azure-openai.md)** guide - - **FireworksAI** ([API keys](https://fireworks.ai/account/api-keys), via custom OpenAI endpoint: https://api.fireworks.ai/inference) - **[OpenRouter](config-openrouter.md)** guide + - **OpenAI-compatible endpoints**: Any provider with an OpenAI-compatible API works out of the box - models, pricing, and capabilities are auto-detected - **Local AI Integrations**: @@ -44,8 +47,9 @@ How to set up AI models and features in big-AGI. - **Enhanced AI Features**: - **[Web Browsing](config-feature-browse.md)**: Enable web page download through third-party services or your own cloud - **Web Search**: Google Search API (see '[Environment Variables](environment-variables.md)') - - **Image Generation**: GPT Image (gpt-image-1), DALL·E 3 and 2 + - **Image Generation**: GPT Image (gpt-image-1), Nano Banana, DALL·E 3 and 2 - **Voice Synthesis**: ElevenLabs, Inworld, OpenAI TTS, LocalAI, or browser Web Speech API + - **[Google Drive](config-feature-google-drive.md)**: Attach files from Google Drive ## Deployment & Customization @@ -64,6 +68,9 @@ For deploying a custom big-AGI instance: - **[Access Control](deploy-authentication.md)**: Optional, add basic user authentication - **[Database Setup](deploy-database.md)**: Optional, enables "Chat Link Sharing" - **[Reverse Proxy](deploy-reverse-proxy.md)**: Optional, enables custom domains and SSL + - **[Docker Deployment](deploy-docker.md)**: Deploy with Docker containers + - **[Kubernetes](deploy-k8s.md)**: Deploy on Kubernetes clusters + - **[Analytics](deploy-analytics.md)**: Set up usage analytics - **[Environment Variables](environment-variables.md)**: Pre-configures models and services ## Community & Support diff --git a/docs/changelog.md b/docs/changelog.md index 0d31628a6..90568b035 100644 --- a/docs/changelog.md +++ b/docs/changelog.md @@ -20,8 +20,11 @@ by release. - And all of the [Big-AGI 2 changes](https://github.com/enricoros/big-AGI/issues/567#issuecomment-2262187617) and more - Built for the future, madly optimized -### What's New in 1.16.1...1.16.9 · Jan 21, 2025 (patch releases) +### What's New in 1.16.1...1.16.13 · (patch releases) +- 1.16.13: Docker fix (#840) +- 1.16.12: Dockerfile update (#840) +- 1.16.11: v1 final release, documentation updates - 1.16.10: OpenRouter models support - 1.16.9: Docker Gemini fix, R1 models support - 1.16.8: OpenAI ChatGPT-4o Latest, o1 models support @@ -70,7 +73,7 @@ by release. - New **[Perplexity](https://www.perplexity.ai/)** and **[Groq](https://groq.com/)** integration (thanks @Penagwin). [#407](https://github.com/enricoros/big-AGI/issues/407), [#427](https://github.com/enricoros/big-AGI/issues/427) - **[LocalAI](https://localai.io/models/)** deep integration, including support for [model galleries](https://github.com/enricoros/big-AGI/issues/411) - **Mistral** Large and Google **Gemini 1.5** support -- Performance optimizations: runs [much faster](https://twitter.com/enricoros/status/1756553038293303434?utm_source=localhost:3000&utm_medium=big-agi), saves lots of power, reduces memory usage +- Performance optimizations: runs [much faster](https://x.com/enricoros/status/1756553038293303434?utm_source=localhost:3000&utm_medium=big-agi), saves lots of power, reduces memory usage - Enhanced UX with auto-sizing charts, refined search and folder functionalities, perfected scaling - And with more UI improvements, documentation, bug fixes (20 tickets), and developer enhancements - [Release notes](https://github.com/enricoros/big-AGI/releases/tag/v1.14.0), and changes [v1.13.1...v1.14.0](https://github.com/enricoros/big-AGI/compare/v1.13.1...v1.14.0) (233 commits, 8,000+ lines changed) diff --git a/docs/config-local-localai.md b/docs/config-local-localai.md index de2e92e4c..2ba3d537d 100644 --- a/docs/config-local-localai.md +++ b/docs/config-local-localai.md @@ -41,6 +41,8 @@ In addition to using the UI, configuration can also be done using ### Integration: Models Gallery +> Note: The Gallery Admin feature described below may have been removed or renamed in recent versions of big-AGI. + If the running LocalAI instance is configured with a [Model Gallery](https://localai.io/models/): - Go to Models > LocalAI diff --git a/docs/config-openrouter.md b/docs/config-openrouter.md index 12623c32e..b13293bbe 100644 --- a/docs/config-openrouter.md +++ b/docs/config-openrouter.md @@ -1,8 +1,7 @@ # OpenRouter Configuration [OpenRouter](https://openrouter.ai) is a standalone, premium service -that provides access to exclusive AI models -such as GPT-4 32k, Claude, and more. These models are typically not available to the public. +that provides access to a wide range of AI models from multiple providers through a single API. This document details the process of integrating OpenRouter with big-AGI. ### 1. OpenRouter Account Setup and API Key Generation @@ -20,7 +19,7 @@ This document details the process of integrating OpenRouter with big-AGI. ![feature-openrouter-add.png](pixels/feature-openrouter-add.png) 3. Input the API key into the **OpenRouter API Key** field, and load the Models. ![feature-openrouter-configure.png](pixels/feature-openrouter-configure.png) -4. OpenAI GPT4-32k and other models will now be accessible and selectable in the application. +4. Models from all supported providers will now be accessible and selectable in the application. In addition to using the UI, configuration can also be done using [environment variables](environment-variables.md). @@ -30,5 +29,5 @@ In addition to using the UI, configuration can also be done using OpenRouter independently manages its service and pricing and is not affiliated with big-AGI. For more detailed information, please visit [this page](https://openrouter.ai/docs#models). -Please note that running large models such as GPT-4 32k can be costly and may rapidly consume -credits - a single prompt may cost $1 or more, at the time of writing. \ No newline at end of file +Please note that running large models can be costly and may rapidly consume credits. +Check model pricing on the OpenRouter website before use. \ No newline at end of file diff --git a/docs/customizations.md b/docs/customizations.md index 180b11943..8a2267ad2 100644 --- a/docs/customizations.md +++ b/docs/customizations.md @@ -49,8 +49,8 @@ Edit the `src/data.ts` file to customize personas. This file houses the default Adapt the UI to match your project's aesthetic, incorporate new features, or exclude unnecessary ones. - [ ] Adjust `src/common/app.theme.ts` for theme changes: colors, spacing, button appearance, animations, etc -- [ ] Modify `src/common/app.config.tsx` to alter the application's name -- [ ] Update `src/common/app.nav.tsx` to revise the navigation bar +- [ ] Modify `src/common/app.release.ts` to alter the application's name +- [ ] Update `src/common/app.nav.ts` to revise the navigation bar ### Add a Message of the Day @@ -71,7 +71,7 @@ Example: `NEXT_PUBLIC_MOTD=🚀 New features available in {{app_build_pkgver}}! Test your application thoroughly using local development (refer to README.md for local build instructions). Deploy using your preferred hosting service. big-AGI supports deployment on platforms like Vercel, Docker, or any Node.js-compatible service, especially those supporting NextJS's "Edge Runtime." -- [deploy-cloudflare.md](deploy-cloudflare.md): for Cloudflare Workers deployment +- [deploy-cloudflare.md](deploy-cloudflare.md): for Cloudflare Pages deployment (limited support) - [deploy-docker.md](deploy-docker.md): for Docker deployment instructions and examples - [deploy-k8s.md](deploy-k8s.md): for Kubernetes deployment instructions and examples diff --git a/docs/deploy-analytics.md b/docs/deploy-analytics.md index 26c2178f1..fd9d76cc8 100644 --- a/docs/deploy-analytics.md +++ b/docs/deploy-analytics.md @@ -51,13 +51,13 @@ Vercel Analytics and Speed Insights are local API endpoints deployed to your dom domain. Furthermore, the Vercel Analytics service is privacy-friendly, and does not track individual users. This service is avaialble to system administrators when deploying to Vercel. It is automatically enabled when deploying to Vercel. -The code that activates Vercel Analytics is located in the `src/pages/_app.tsx` file: +The code that activates Vercel Analytics is located in the `pages/_app.tsx` file: ```tsx const MyApp = ({ Component, emotionCache, pageProps }: MyAppProps) => <> ... - {isVercelFromFrontend && } - {isVercelFromFrontend && } + {Is.Deployment.VercelFromFrontend && } + {Is.Deployment.VercelFromFrontend && } ... ; ``` diff --git a/docs/deploy-cloudflare.md b/docs/deploy-cloudflare.md index c93075d25..0294934fa 100644 --- a/docs/deploy-cloudflare.md +++ b/docs/deploy-cloudflare.md @@ -1,18 +1,20 @@ +--- +unlisted: true +--- + # Deploying a Next.js App on Cloudflare Pages -> WARNING: Cloudflare Pages does not support traditional NodeJS runtimes, but only Edge Runtime functions. +> WARNING: Cloudflare Pages only supports Edge Runtime functions, not the full Node.js runtime. > -> In this project we use Prisma connected to serverless Postgres, which at the moment cannot run on -> edge functions, so we cannot deploy this project on Cloudflare Pages. +> The cloud router in this project requires a Node.js runtime for Supabase SDK, authentication, +> sync, and other server-side features that cannot run on Cloudflare's edge runtime. > -> Workaround: Step 3.4. has been added below, to DELETE the NodeJS traditional runtime - which means that some +> Workaround: Step 3.4. has been added below, to DELETE the Node.js cloud router - which means that some > parts of this application will not work. -> - [Side effects](https://github.com/enricoros/big-agi/blob/main/src/apps/chat/trade/server/trade.router.ts#L19): - > Sharing functionality to DB, and import from ChatGPT share, and post to Paste.GG will not work +> - [Side effects](https://github.com/enricoros/big-agi/blob/main/src/modules/trade/server/trade.router.ts): +> Sharing functionality, import from ChatGPT share, and post to Paste.GG will not work +> - Cloud features (sync, auth, payments) will not be available > - See [Issue 174](https://github.com/enricoros/big-agi/issues/174). -> -> Longer term: follow [prisma/prisma: Support Edge Function deployments](https://github.com/prisma/prisma/issues/21394) -> and convert the Node runtime to Edge runtime once Prisma supports it. This guide provides steps to deploy your Next.js app on Cloudflare Pages. It is based on the [official Cloudflare developer documentation](https://developers.cloudflare.com/pages/framework-guides/deploy-a-nextjs-site/), diff --git a/docs/draft-big-agi.md b/docs/draft-big-agi.md deleted file mode 100644 index 733ddff9f..000000000 --- a/docs/draft-big-agi.md +++ /dev/null @@ -1,14 +0,0 @@ -# Why big-AGI? -Placeholder for a document that demonstrates the productivity and unique features of Big-AGI. - -## Exclusive features -- [x] Call AGI -- [x] Continuous Voice mode -- [x] Diagram generation -- [ ] ... - -## Productivity Features -- [x] Multi-window to never wait -- [x] Multi-Chat to explore different solutions -- [x] Rendering of graphs, charts, mindmaps -- [ ] ... \ No newline at end of file diff --git a/docs/help-advanced-tricks.md b/docs/help-advanced-tricks.md index 8fe73f7c0..18a0b5c0d 100644 --- a/docs/help-advanced-tricks.md +++ b/docs/help-advanced-tricks.md @@ -1,3 +1,7 @@ +--- +unlisted: true +--- + # Big-AGI Advanced Tips & Tricks > 🚨 This file is not meant for publication, and it's just been created as a handbook with tips diff --git a/docs/help-data-ownership.md b/docs/help-data-ownership.md index 606dde2ef..bfa124502 100644 --- a/docs/help-data-ownership.md +++ b/docs/help-data-ownership.md @@ -30,6 +30,12 @@ You can see your data in your browser's local storage and IndexedDB - try it you ![Browser local storage showing API keys and chat data](pixels/data_ownership_local_storage.png) +### Sync for Authenticated Users + +Users with accounts on big-agi.com who opt into Sync (a Pro feature) have their entity data - such as conversations and personas - replicated to the server for multi-device access. +Server-side data is isolated per-user using Row Level Security (RLS), ensuring that no other user can access your synced data. +Sync is entirely optional; without it, all data remains local to your browser. + ### What This Means For You Storing data in your browser means: @@ -43,7 +49,7 @@ Storing data in your browser means: Big-AGI generates a _device identifier_ that combines timestamp and random components, stored only on your device. This identifier: -- Is used only for the **optional sync functionality** between your devices (not yet ready) +- Is used only for the **optional sync functionality** between your devices - Helps maintain data consistency when using Big-AGI across multiple devices - Remains completely local unless you explicitly enable sync - Is not used for tracking, analytics, or telemetry diff --git a/docs/installation.md b/docs/installation.md index 0af4525ef..6dc50c560 100644 --- a/docs/installation.md +++ b/docs/installation.md @@ -7,7 +7,7 @@ process for your own instance of big-AGI and related products. **Try big-AGI** - You don't need to install anything if you want to play with big-AGI and have your API keys to various model services. You can access our free instance on [big-AGI.com](https://big-agi.com). -The free instance runs the latest `main-stable` branch from this repository. +The free instance runs the latest `main` branch from this repository. ## 🧩 Build-your-own @@ -72,9 +72,8 @@ Create your GitHub fork, create a Vercel project over that fork, and deploy it. ### Deploy on Cloudflare -Deploy on Cloudflare's global network by installing big-AGI on -Cloudflare Pages. Check out the [Cloudflare Installation Guide](deploy-cloudflare.md) -for step-by-step instructions. +> Note: Cloudflare Pages deployment has limitations due to Edge Runtime constraints. +> See the [Cloudflare guide](deploy-cloudflare.md) for details and known issues. ### Docker Deployments @@ -146,6 +145,6 @@ Enjoy all the features of big-AGI without the hassle of infrastructure managemen Join our vibrant community of developers, researchers, and AI enthusiasts. Share your projects, get help, and collaborate with others. - [Discord Community](https://discord.gg/MkH4qj2Jp9) -- [Twitter](https://twitter.com/enricoros) +- [X (Twitter)](https://x.com/enricoros) For any questions or inquiries, please don't hesitate to [reach out to our team](mailto:hello@big-agi.com). diff --git a/docs/use-chat-react.md b/docs/use-chat-react.md index 194d8647d..f8031527b 100644 --- a/docs/use-chat-react.md +++ b/docs/use-chat-react.md @@ -1,3 +1,7 @@ +--- +unlisted: true +--- + # ReAct: question answering with Reasoning and Actions ## What is ReAct? diff --git a/kb/modules/AIX-callers-analysis.md b/kb/modules/AIX-callers-analysis.md index f86fcdc24..322665947 100644 --- a/kb/modules/AIX-callers-analysis.md +++ b/kb/modules/AIX-callers-analysis.md @@ -7,8 +7,8 @@ This document analyzes all AIX function callers and their patterns for message r ### Three-Tier Call Hierarchy **Core AIX Functions** (Direct tRPC API callers): -- `aixChatGenerateContent_DMessage_FromConversation` - 8 callers (conversation streaming) -- `aixChatGenerateContent_DMessage` - 6 callers (direct request/response) +- `aixChatGenerateContent_DMessage_FromConversation` - 9 callers (conversation streaming) +- `aixChatGenerateContent_DMessage_orThrow` - 6 callers (direct request/response) - `aixChatGenerateText_Simple` - 12 callers (text-only utilities) **Utility Layer** (Hooks & Functions): @@ -24,6 +24,7 @@ This document analyzes all AIX function callers and their patterns for message r | **Caller** | **Context** | **Message Removal** | **Placeholder** | **Error Handling** | |------------|-------------|-------------------|----------------|-------------------| | **Chat Persona** | `'conversation'` | `messageWasInterruptedAtStart()` → `removeMessage()` | None | Error fragments | +| **XE Chat Generate** | `'conversation'` | `messageWasInterruptedAtStart()` → `removeMessage()` | `'...'` placeholder | Error fragments via messageEditor | | **Beam Scatter** | `'beam-scatter'` | `messageWasInterruptedAtStart()` → empty message | `SCATTER_PLACEHOLDER` | Ray status update | | **Beam Gather** | `'beam-gather'` | `messageWasInterruptedAtStart()` → clear fragments | `GATHER_PLACEHOLDER` | Re-throw errors | | **Beam Follow-up** | `'beam-followup'` | `messageWasInterruptedAtStart()` → remove message | `FOLLOWUP_PLACEHOLDER` | Status updates | diff --git a/kb/modules/AIX.md b/kb/modules/AIX.md index ad8ac3627..3522afa6f 100644 --- a/kb/modules/AIX.md +++ b/kb/modules/AIX.md @@ -92,12 +92,12 @@ AIX is organized into the following files and folders: - Dispatch (`/server/dispatch/`) - Server to AI Provider communication: - `/server/dispatch/chatGenerate/`: Content Generation with chat-style inputs: - - `./adapters/`: Adapters for creating API requests for different AI protocols (Anthropic, Gemini, OpenAI). - - `./parsers/`: Parsers for parsing streaming/non-streamin responses from different AI protocols (same 3). + - `./adapters/`: Adapters for creating API requests for different AI protocols (Anthropic, Bedrock, Gemini, OpenAI Chat Completions, OpenAI Responses, xAI Responses). + - `./parsers/`: Parsers for parsing streaming/non-streaming responses from different AI protocols (Anthropic, Bedrock Converse, Gemini, OpenAI, OpenAI Responses). - `chatGenerate.dispatch.ts`: Creates a pipeline to execute Chat Generation to a specific provider. - `ChatGenerateTransmitter.ts`: Used to serialize and transmit AixWire_Particles to the client. - `/server/dispatch/wiretypes/`: AI provider Wire Types: - - Type definitions for different AI providers/protocols (Anthropic, Gemini, OpenAI). + - Type definitions for different AI providers/protocols (Anthropic, Bedrock Converse, Gemini, OpenAI, xAI). - `stream.demuxers.ts`: Handles demuxing of different stream formats. ## 3. Architecture Diagram @@ -160,7 +160,7 @@ sequenceDiagram AIX Client ->> AIX Client: Display error message else DMessageDocPart AIX Client ->> AIX Client: Process and display document - else DMetaPlaceholderPart + else DVoidPlaceholderPart AIX Client ->> AIX Client: Handle placeholder (non-submitted) end end diff --git a/kb/systems/LLM-parameters-system.md b/kb/systems/LLM-parameters-system.md index 54797e4da..4573d7b64 100644 --- a/kb/systems/LLM-parameters-system.md +++ b/kb/systems/LLM-parameters-system.md @@ -47,7 +47,7 @@ Shows only parameters that are: - Not marked as `hidden` **Value Resolution**: Both UIs use `getAllModelParameterValues()` to merge: -1. **Fallback values** - Implicit parameters get their `runtimeFallback` values +1. **Fallback values** - Implicit parameters get their `LLMImplicitParametersRuntimeFallback` values 2. **Initial values** - Model's `initialParameters` (populated during model creation) 3. **User values** - User's `userParameters` (highest priority) @@ -63,7 +63,7 @@ The AIX client transforms DLLM parameters to wire protocol format. This layer ha Server-side adapters translate AIX parameters to vendor APIs. Each vendor may interpret parameters differently: -- **OpenAI**: `vndEffort` → `reasoning_effort` +- **OpenAI**: `vndEffort` -> `reasoning_effort` - **Perplexity**: Reuses OpenAI parameter format - **OpenAI Responses API**: Maps to structured reasoning config with additional logic @@ -71,7 +71,7 @@ Server-side adapters translate AIX parameters to vendor APIs. Each vendor may in When a model is loaded: -1. **Model Creation**: `modelDescriptionToDLLM()` creates the DLLM with empty `initialParameters` +1. **Model Creation**: `_createDLLMFromModelDescription()` creates the DLLM with empty `initialParameters` 2. **Initial Value Application**: `applyModelParameterSpecsInitialValues()` populates initial values from: - Model spec `initialValue` (highest priority) - Registry `initialValue` (fallback) @@ -117,4 +117,4 @@ The architecture supports parameter evolution: - **UI Controls**: `src/modules/llms/models-modal/LLMParametersEditor.tsx` - **AIX Translation**: `src/modules/aix/client/aix.client.ts` - **Wire Types**: `src/modules/aix/server/api/aix.wiretypes.ts` -- **Vendor Adapters**: `src/modules/aix/server/dispatch/chatGenerate/adapters/*.ts` \ No newline at end of file +- **Vendor Adapters**: `src/modules/aix/server/dispatch/chatGenerate/adapters/*.ts` diff --git a/kb/systems/client-side-fetch.md b/kb/systems/client-side-fetch.md index 1fd6934c0..221f12f78 100644 --- a/kb/systems/client-side-fetch.md +++ b/kb/systems/client-side-fetch.md @@ -6,7 +6,7 @@ Client-Side Fetch (CSF) enables direct browser-to-API communication, bypassing t CSF is implemented as an opt-in setting stored as `csf: boolean` in each vendor's service settings. The vendor interface exposes `csfAvailable?: (setup) => boolean` to determine if CSF can be enabled (typically checking if an API key or host is configured). The actual execution happens in `aix.client.direct-chatGenerate.ts` which dynamically imports when CSF is active, making direct fetch calls using the same wire protocols as the server. -All 17 supported vendors (OpenAI, Anthropic, Gemini, Ollama, LocalAI, Deepseek, Groq, Mistral, xAI, OpenRouter, Perplexity, Together AI, Alibaba, Moonshot, OpenPipe, LM Studio, Z.ai) support CSF. Cloud vendors require CORS support from the API provider (all tested vendors return `access-control-allow-origin: *`). Local vendors (Ollama, LocalAI, LM Studio) require CORS to be enabled on the local server. +All 20+ supported vendors (OpenAI, Anthropic, Gemini, Ollama, LocalAI, Deepseek, Groq, Mistral, xAI, OpenRouter, Perplexity, Together AI, Alibaba, Moonshot, OpenPipe, LM Studio, Z.ai, Azure, Bedrock) support CSF. Cloud vendors require CORS support from the API provider (all tested vendors return `access-control-allow-origin: *`). Local vendors (Ollama, LocalAI, LM Studio) require CORS to be enabled on the local server. ## UI diff --git a/package.json b/package.json index 65b265c62..628ad43a2 100644 --- a/package.json +++ b/package.json @@ -2,7 +2,8 @@ "name": "big-agi", "version": "2.0.3", "private": true, - "author": "Enrico Ros ", + "author": "Enrico Ros (https://www.enricoros.com)", + "homepage": "https://big-agi.com", "repository": "https://github.com/enricoros/big-agi", "scripts": { "dev": "next dev --turbopack", diff --git a/pages/_document.tsx b/pages/_document.tsx index 158b9399e..6cb6859c0 100644 --- a/pages/_document.tsx +++ b/pages/_document.tsx @@ -37,14 +37,18 @@ export default function MyDocument({ emotionStyleTags }: MyDocumentProps) { - {/* Twitter */} - + {/* Twitter / X */} + {Brand.URIs.CardImage && } - + + + {/* Author */} + + {/* Style Sheets (injected and server-side) */} diff --git a/src/apps/news/news.data.tsx b/src/apps/news/news.data.tsx index 6de42da79..db24b99ac 100644 --- a/src/apps/news/news.data.tsx +++ b/src/apps/news/news.data.tsx @@ -174,7 +174,7 @@ export const NewsItems: NewsItem[] = [ { text: <>Support for new Mistral-Large models, icon: MistralIcon }, { text: <>Support for Google Gemini 1.5 models and various improvements, icon: GoogleIcon as any }, { text: <>Deeper LocalAI integration including support for model galleries, icon: LocalAIIcon }, - { text: <>Major performance optimizations: runs faster, saves power, saves memory }, + { text: <>Major performance optimizations: runs faster, saves power, saves memory }, { text: <>Improvements: auto-size charts, search and folder experience }, { text: <>Perfect chat scaling, with rapid keyboard shortcuts }, { text: <>Also: diagrams auto-resize, open code with StackBlitz and JSFiddle, quick model visibility toggle, open links externally, docs on the web },