diff --git a/README.md b/README.md index b1ec83b4..8a5d47e1 100644 --- a/README.md +++ b/README.md @@ -74,19 +74,19 @@ | Talk Interface | Generate Audio | | --- | --- | -| ![Screenshot 2025-03-31 at 12-01-36 LocalAI - Talk](https://github.com/user-attachments/assets/9841b1ee-88af-4b96-8ec0-41b17364efa7) | ![Screenshot 2025-03-31 at 12-01-29 LocalAI - Generate audio with voice-en-us-ryan-low](https://github.com/user-attachments/assets/d729f6f4-0621-4715-bda3-35fe6e159524) | +| ![Screenshot 2025-03-31 at 12-01-36 LocalAI - Talk](./docs/assets/images/screenshots/screenshot_tts.png) | ![Screenshot 2025-03-31 at 12-01-29 LocalAI - Generate audio with voice-en-us-ryan-low](./docs/assets/images/screenshots/screenshot_tts.png) | | Models Overview | Generate Images | | --- | --- | -| ![Screenshot 2025-03-31 at 12-01-20 LocalAI - Models](https://github.com/user-attachments/assets/3cf0b918-ba8e-498a-a3cd-485db5984325) | ![Screenshot 2025-03-31 at 12-31-41 LocalAI - Generate images with flux 1-dev](https://github.com/user-attachments/assets/6753d23d-218b-4e07-94b8-9e6c5a4f2311) | +| ![Screenshot 2025-03-31 at 12-01-20 LocalAI - Models](./docs/assets/images/screenshots/screenshot_gallery.png) | ![Screenshot 2025-03-31 at 12-31-41 LocalAI - Generate images with flux 1-dev](./docs/assets/images/screenshots/screenshot_image.png) | -| Chat Interface | API Overview | +| Chat Interface | Home | | --- | --- | -| ![Screenshot 2025-03-31 at 11-57-44 LocalAI - Chat with localai-functioncall-qwen2 5-7b-v0 5](https://github.com/user-attachments/assets/048eab31-0f0c-4d52-a920-3715233f9bf3) | ![Screenshot 2025-03-31 at 11-57-23 LocalAI API - c2a39e3 (c2a39e3639227cfd94ffffe9f5691239acc275a8)](https://github.com/user-attachments/assets/2540e8ce-1a2c-4c12-800c-763bd9be247f) | +| ![Screenshot 2025-03-31 at 11-57-44 LocalAI - Chat with localai-functioncall-qwen2 5-7b-v0 5](./docs/assets/images/screenshots/screenshot_chat.png) | ![Screenshot 2025-03-31 at 11-57-23 LocalAI API - c2a39e3 (c2a39e3639227cfd94ffffe9f5691239acc275a8)](./docs/assets/images/screenshots/screenshot_home.png) | | Login | Swarm | | --- | --- | -|![Screenshot 2025-03-31 at 12-09-59 ](https://github.com/user-attachments/assets/5af681b0-dd8e-4fe8-a234-a22f8a040547) | ![Screenshot 2025-03-31 at 12-10-39 LocalAI - P2P dashboard](https://github.com/user-attachments/assets/b9527176-63d6-4d2e-8ed1-7fde13a9b0ad) | +|![Screenshot 2025-03-31 at 12-09-59 ](./docs/assets/images/screenshots/screenshot_login.png) | ![Screenshot 2025-03-31 at 12-10-39 LocalAI - P2P dashboard](./docs/assets/images/screenshots/screenshot_p2p.png) | ## 💻 Quickstart diff --git a/docs/assets/images/imagen.png b/docs/assets/images/imagen.png new file mode 100644 index 00000000..7d9808f4 Binary files /dev/null and b/docs/assets/images/imagen.png differ diff --git a/docs/assets/images/logos/logo.svg b/docs/assets/images/logos/logo.svg new file mode 100644 index 00000000..5e881d4b --- /dev/null +++ b/docs/assets/images/logos/logo.svg @@ -0,0 +1,171 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + diff --git a/docs/assets/images/screenshots/screenshot_chat.png b/docs/assets/images/screenshots/screenshot_chat.png new file mode 100644 index 00000000..bc621ba7 Binary files /dev/null and b/docs/assets/images/screenshots/screenshot_chat.png differ diff --git a/docs/assets/images/screenshots/screenshot_gallery.png b/docs/assets/images/screenshots/screenshot_gallery.png new file mode 100644 index 00000000..c8a33642 Binary files /dev/null and b/docs/assets/images/screenshots/screenshot_gallery.png differ diff --git a/docs/assets/images/screenshots/screenshot_home.png b/docs/assets/images/screenshots/screenshot_home.png new file mode 100644 index 00000000..18777a46 Binary files /dev/null and b/docs/assets/images/screenshots/screenshot_home.png differ diff --git a/docs/assets/images/screenshots/screenshot_image.png b/docs/assets/images/screenshots/screenshot_image.png new file mode 100644 index 00000000..7d9808f4 Binary files /dev/null and b/docs/assets/images/screenshots/screenshot_image.png differ diff --git a/docs/assets/images/screenshots/screenshot_login.png b/docs/assets/images/screenshots/screenshot_login.png new file mode 100644 index 00000000..82b1614a Binary files /dev/null and b/docs/assets/images/screenshots/screenshot_login.png differ diff --git a/docs/assets/images/screenshots/screenshot_p2p.png b/docs/assets/images/screenshots/screenshot_p2p.png new file mode 100644 index 00000000..fbeb75ef Binary files /dev/null and b/docs/assets/images/screenshots/screenshot_p2p.png differ diff --git a/docs/assets/images/screenshots/screenshot_talk.png b/docs/assets/images/screenshots/screenshot_talk.png new file mode 100644 index 00000000..956b59d1 Binary files /dev/null and b/docs/assets/images/screenshots/screenshot_talk.png differ diff --git a/docs/assets/images/screenshots/screenshot_tts.png b/docs/assets/images/screenshots/screenshot_tts.png new file mode 100644 index 00000000..8df68f70 Binary files /dev/null and b/docs/assets/images/screenshots/screenshot_tts.png differ diff --git a/docs/config.toml b/docs/config.toml index dafaa6d6..af583c59 100644 --- a/docs/config.toml +++ b/docs/config.toml @@ -50,7 +50,7 @@ defaultContentLanguage = 'en' logo = "https://raw.githubusercontent.com/mudler/LocalAI/53652d5b8f8049e95ef71f0fed39671eb39c211c/core/http/static/logo.png" logo_text = "" - title = "LocalAI documentation" # default html title for documentation pages/sections + title = "LocalAI" # default html title for documentation pages/sections pathName = "docs" # path name for documentation site | default "docs" diff --git a/docs/content/docs/features/distributed_inferencing.md b/docs/content/docs/features/distributed_inferencing.md index 71d29f39..d599de87 100644 --- a/docs/content/docs/features/distributed_inferencing.md +++ b/docs/content/docs/features/distributed_inferencing.md @@ -13,6 +13,8 @@ LocalAI supports two modes of distributed inferencing via p2p: - **Federated Mode**: Requests are shared between the cluster and routed to a single worker node in the network based on the load balancer's decision. - **Worker Mode** (aka "model sharding" or "splitting weights"): Requests are processed by all the workers which contributes to the final inference result (by sharing the model weights). +A list of global instances shared by the community is available at [explorer.localai.io](https://explorer.localai.io). + ## Usage Starting LocalAI with `--p2p` generates a shared token for connecting multiple instances: and that's all you need to create AI clusters, eliminating the need for intricate network setups. diff --git a/docs/content/docs/getting-started/quickstart.md b/docs/content/docs/getting-started/quickstart.md index 4e14c505..0d962d3c 100644 --- a/docs/content/docs/getting-started/quickstart.md +++ b/docs/content/docs/getting-started/quickstart.md @@ -18,14 +18,45 @@ To access the WebUI with an API_KEY, browser extensions such as [Requestly](http {{% /alert %}} -## Using the Bash Installer +## Quickstart -Install LocalAI easily using the bash installer with the following command: -```sh +### Using the Bash Installer +```bash curl https://localai.io/install.sh | sh ``` +### Run with docker: +```bash +# CPU only image: +docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-cpu + +# Nvidia GPU: +docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gpu-nvidia-cuda-12 + +# CPU and GPU image (bigger size): +docker run -ti --name local-ai -p 8080:8080 localai/localai:latest + +# AIO images (it will pre-download a set of models ready for use, see https://localai.io/basics/container/) +docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-aio-cpu +``` + +### Load models: + +```bash +# From the model gallery (see available models with `local-ai models list`, in the WebUI from the model tab, or visiting https://models.localai.io) +local-ai run llama-3.2-1b-instruct:q4_k_m +# Start LocalAI with the phi-2 model directly from huggingface +local-ai run huggingface://TheBloke/phi-2-GGUF/phi-2.Q8_0.gguf +# Install and run a model from the Ollama OCI registry +local-ai run ollama://gemma:2b +# Run a model from a configuration file +local-ai run https://gist.githubusercontent.com/.../phi-2.yaml +# Install and run a model from a standard OCI registry (e.g., Docker Hub) +local-ai run oci://localai/phi-2:latest +``` + + For a full list of options, refer to the [Installer Options]({{% relref "docs/advanced/installer" %}}) documentation. Binaries can also be [manually downloaded]({{% relref "docs/reference/binaries" %}}). diff --git a/docs/content/docs/overview.md b/docs/content/docs/overview.md index 16b82a31..981ba765 100644 --- a/docs/content/docs/overview.md +++ b/docs/content/docs/overview.md @@ -1,4 +1,3 @@ - +++ title = "Overview" weight = 1 @@ -10,157 +9,93 @@ author = "Ettore Di Giacinto" icon = "info" +++ -

- -

+# Welcome to LocalAI -

- -LocalAI forks - - -LocalAI stars - - -LocalAI pull-requests - - - - -

+LocalAI is your complete AI stack for running AI models locally. It's designed to be simple, efficient, and accessible, providing a drop-in replacement for OpenAI's API while keeping your data private and secure. -

- -LocalAI Docker hub - - -LocalAI Quay.io - -

+## Why LocalAI? -

-mudler%2FLocalAI | Trendshift -

+In today's AI landscape, privacy, control, and flexibility are paramount. LocalAI addresses these needs by: -

- -Follow LocalAI_API - - -Join LocalAI Discord Community - -

+- **Privacy First**: Your data never leaves your machine +- **Complete Control**: Run models on your terms, with your hardware +- **Open Source**: MIT licensed and community-driven +- **Flexible Deployment**: From laptops to servers, with or without GPUs +- **Extensible**: Add new models and features as needed -> 💡 Get help - [❓FAQ](https://localai.io/faq/) [💭Discussions](https://github.com/go-skynet/LocalAI/discussions) [💭Discord](https://discord.gg/uJAeKSAGDy) -> -> [💻 Quickstart](https://localai.io/basics/getting_started/) [🖼️ Models](https://models.localai.io/) [🚀 Roadmap](https://github.com/mudler/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3Aroadmap) [🥽 Demo](https://demo.localai.io) [🌍 Explorer](https://explorer.localai.io) [🛫 Examples](https://github.com/go-skynet/LocalAI/tree/master/examples/) +## Core Components +LocalAI is more than just a single tool - it's a complete ecosystem: -**LocalAI** is the free, Open Source OpenAI alternative. LocalAI act as a drop-in replacement REST API that's compatible with OpenAI API specifications for local inferencing. It allows you to run LLMs, generate images, audio (and not only) locally or on-prem with consumer grade hardware, supporting multiple model families and architectures. Does not require GPU. It is created and maintained by [Ettore Di Giacinto](https://github.com/mudler). +1. **[LocalAI Core](https://github.com/mudler/LocalAI)** + - OpenAI-compatible API + - Multiple model support (LLMs, image, audio) + - No GPU required + - Fast inference with native bindings + - [Github repository](https://github.com/mudler/LocalAI) +2. **[LocalAGI](https://github.com/mudler/LocalAGI)** + - Autonomous AI agents + - No coding required + - WebUI and REST API support + - Extensible agent framework + - [Github repository](https://github.com/mudler/LocalAGI) -## Start LocalAI +3. **[LocalRecall](https://github.com/mudler/LocalRecall)** + - Semantic search + - Memory management + - Vector database + - Perfect for AI applications + - [Github repository](https://github.com/mudler/LocalRecall) -Start the image with Docker to have a functional clone of OpenAI! 🚀: +## Getting Started -```bash -docker run -p 8080:8080 --name local-ai -ti localai/localai:latest-aio-cpu -# Do you have a Nvidia GPUs? Use this instead -# CUDA 11 -# docker run -p 8080:8080 --gpus all --name local-ai -ti localai/localai:latest-aio-gpu-nvidia-cuda-11 -# CUDA 12 -# docker run -p 8080:8080 --gpus all --name local-ai -ti localai/localai:latest-aio-gpu-nvidia-cuda-12 -``` - -Or just use the bash installer: +The fastest way to get started is with our one-line installer: ```bash curl https://localai.io/install.sh | sh ``` -See the [💻 Quickstart](https://localai.io/basics/getting_started/) for all the options and way you can run LocalAI! +Or use Docker for a quick start: -## What is LocalAI? +```bash +docker run -p 8080:8080 --name local-ai -ti localai/localai:latest-aio-cpu +``` -In a nutshell: +For more detailed installation options and configurations, see our [Getting Started guide](/basics/getting_started/). -- Local, OpenAI drop-in alternative REST API. You own your data. -- NO GPU required. NO Internet access is required either - - Optional, GPU Acceleration is available. See also the [build section](https://localai.io/basics/build/index.html). -- Supports multiple models -- 🏃 Once loaded the first time, it keep models loaded in memory for faster inference -- ⚡ Doesn't shell-out, but uses bindings for a faster inference and better performance. +## Key Features -LocalAI is focused on making the AI accessible to anyone. Any contribution, feedback and PR is welcome! +- **Text Generation**: Run various LLMs locally +- **Image Generation**: Create images with stable diffusion +- **Audio Processing**: Text-to-speech and speech-to-text +- **Vision API**: Image understanding and analysis +- **Embeddings**: Vector database support +- **Functions**: OpenAI-compatible function calling +- **P2P**: Distributed inference capabilities -Note that this started just as a fun weekend project by [mudler](https://github.com/mudler) in order to try to create the necessary pieces for a full AI assistant like `ChatGPT`: the community is growing fast and we are working hard to make it better and more stable. If you want to help, please consider contributing (see below)! +## Community and Support -### 🚀 Features +LocalAI is a community-driven project. You can: -- 📖 [Text generation with GPTs](https://localai.io/features/text-generation/) (`llama.cpp`, `gpt4all.cpp`, ... [:book: and more](https://localai.io/model-compatibility/index.html#model-compatibility-table)) -- 🗣 [Text to Audio](https://localai.io/features/text-to-audio/) -- 🔈 [Audio to Text](https://localai.io/features/audio-to-text/) (Audio transcription with `whisper.cpp`) -- 🎨 [Image generation with stable diffusion](https://localai.io/features/image-generation) -- 🔥 [OpenAI functions](https://localai.io/features/openai-functions/) 🆕 -- 🧠 [Embeddings generation for vector databases](https://localai.io/features/embeddings/) -- ✍️ [Constrained grammars](https://localai.io/features/constrained_grammars/) -- 🖼️ [Download Models directly from Huggingface ](https://localai.io/models/) -- 🥽 [Vision API](https://localai.io/features/gpt-vision/) -- 💾 [Stores](https://localai.io/stores) -- 📈 [Reranker](https://localai.io/features/reranker/) -- 🆕🖧 [P2P Inferencing](https://localai.io/features/distribute/) +- Join our [Discord community](https://discord.gg/uJAeKSAGDy) +- Check out our [GitHub repository](https://github.com/mudler/LocalAI) +- Contribute to the project +- Share your use cases and examples -## Contribute and help +## Next Steps -To help the project you can: +Ready to dive in? Here are some recommended next steps: -- If you have technological skills and want to contribute to development, have a look at the open issues. If you are new you can have a look at the [good-first-issue](https://github.com/go-skynet/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3A%22good+first+issue%22) and [help-wanted](https://github.com/go-skynet/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3A%22help+wanted%22) labels. +1. [Install LocalAI](/basics/getting_started/) +2. [Explore available models](https://models.localai.io) +3. [Model compatibility](/model-compatibility/) +4. [Try out examples](https://github.com/mudler/LocalAI-examples) +5. [Join the community](https://discord.gg/uJAeKSAGDy) +6. [Check the LocalAI Github repository](https://github.com/mudler/LocalAI) +7. [Check the LocalAGI Github repository](https://github.com/mudler/LocalAGI) -- If you don't have technological skills you can still help improving documentation or [add examples](https://github.com/go-skynet/LocalAI/tree/master/examples) or share your user-stories with our community, any help and contribution is welcome! -## 🌟 Star history +## License -[![LocalAI Star history Chart](https://api.star-history.com/svg?repos=mudler/LocalAI&type=Date)](https://star-history.com/#mudler/LocalAI&Date) - -## ❤️ Sponsors - -> Do you find LocalAI useful? - -Support the project by becoming [a backer or sponsor](https://github.com/sponsors/mudler). Your logo will show up here with a link to your website. - -A huge thank you to our generous sponsors who support this project covering CI expenses, and our [Sponsor list](https://github.com/sponsors/mudler): - -

- - - - -
-
-

- -## 📖 License - -LocalAI is a community-driven project created by [Ettore Di Giacinto](https://github.com/mudler/). - -MIT - Author Ettore Di Giacinto - -## 🙇 Acknowledgements - -LocalAI couldn't have been built without the help of great software already available from the community. Thank you! - -- [llama.cpp](https://github.com/ggerganov/llama.cpp) -- https://github.com/tatsu-lab/stanford_alpaca -- https://github.com/cornelk/llama-go for the initial ideas -- https://github.com/antimatter15/alpaca.cpp -- https://github.com/EdVince/Stable-Diffusion-NCNN -- https://github.com/ggerganov/whisper.cpp -- https://github.com/saharNooby/rwkv.cpp -- https://github.com/rhasspy/piper - -## 🤗 Contributors - -This is a community project, a special thanks to our contributors! 🤗 - - - +LocalAI is MIT licensed, created and maintained by [Ettore Di Giacinto](https://github.com/mudler). diff --git a/docs/data/landing.yaml b/docs/data/landing.yaml index cb40af53..ff3f02e6 100644 --- a/docs/data/landing.yaml +++ b/docs/data/landing.yaml @@ -25,7 +25,10 @@ hero: height: 540px title: "" - subtitle: "**The free, OpenAI, Anthropic, et al alternative. Your All-in-One Complete AI Stack** - Run powerful language models, autonomous agents, and document intelligence **locally** on your hardware. **No cloud, no limits, no compromise.**" + subtitle: | + **The free, OpenAI, Anthropic alternative. Your All-in-One Complete AI Stack** - Run powerful language models, autonomous agents, and document intelligence **locally** on your hardware. + + **No cloud, no limits, no compromise.** image: path: "images" @@ -44,7 +47,10 @@ hero: url: "https://github.com/mudler/LocalAI" info: | - **Drop-in replacement for OpenAI API** - modular suite of tools that work seamlessly together or independently. Start with LocalAI's OpenAI-compatible API, extend with autonomous agents, and enhance with semantic search - all running locally on your hardware. + **Drop-in replacement for OpenAI API** - modular suite of tools that work seamlessly together or independently. + + Start with **[LocalAI](https://localai.io)**'s OpenAI-compatible API, extend with **[LocalAGI](https://github.com/mudler/LocalAGI)**'s autonomous agents, and enhance with **[LocalRecall](https://github.com/mudler/LocalRecall)**'s semantic search - all running locally on your hardware. + **Open Source** MIT Licensed. # Feature Grid @@ -59,21 +65,22 @@ featureGrid: items: - title: LLM Inferencing - icon: api + icon: memory_alt description: LocalAI is a free, **Open Source** OpenAI alternative. Run **LLMs**, generate **images**, **audio** and more **locally** with consumer grade hardware. ctaLink: text: learn more url: /basics/getting_started/ - title: Agentic-first - icon: ai + icon: smart_toy description: | - Extend LocalAI with LocalAGI, an autonomous AI agent platform that runs locally. Build and deploy autonomous agents with ease. Interact with REST APIs or use the WebUI. + Extend LocalAI with LocalAGI, an autonomous AI agent platform that runs locally, no coding required. + Build and deploy autonomous agents with ease. Interact with REST APIs or use the WebUI. ctaLink: text: learn more url: https://github.com/mudler/LocalAGI - title: Memory and Knowledge base - icon: memory + icon: psychology description: Extend LocalAI with LocalRecall, A local rest api for semantic search and memory management. Perfect for AI applications. ctaLink: @@ -99,21 +106,21 @@ featureGrid: description: | Support for various model families including LLMs, image generation, and audio models. Supports multiple backends for inferencing, including vLLM, llama.cpp, and more. - You can switch between them as needed. + You can switch between them as needed and install them from the Web interface or the CLI. ctaLink: text: learn more url: /model-compatibility - title: Privacy Focused - icon: lock + icon: security description: Keep your data local. No data leaves your machine, ensuring complete privacy. ctaLink: text: learn more - url: /basics/getting_started/ + url: /basics/container/ - title: Easy Setup icon: settings - description: Simple installation and configuration. Get started in minutes with Docker or local installation. + description: Simple installation and configuration. Get started in minutes with Binaries installation, Docker, Podman, Kubernetes or local installation. ctaLink: text: learn more url: /basics/getting_started/ @@ -125,12 +132,7 @@ featureGrid: text: learn more url: https://github.com/mudler/LocalAI - - title: Flexible Deployment - icon: deploy - description: Deploy anywhere - from your laptop to your server. Supports Docker, Kubernetes, and more. - ctaLink: - text: learn more - url: /basics/container/ + - title: Extensible icon: extension @@ -140,7 +142,7 @@ featureGrid: url: /docs/integrations/ - title: Peer 2 Peer - icon: network + icon: hub description: | LocalAI is designed to be a decentralized LLM inference, powered by a peer-to-peer system based on libp2p. It is designed to be used in a local or remote network, and is compatible with any LLM model. @@ -162,7 +164,8 @@ imageText: template: image text title: Run AI models locally with ease - subtitle: LocalAI makes it simple to run various AI models on your own hardware, from text generation to image creation. + subtitle: | + LocalAI makes it simple to run various AI models on your own hardware. From text generation to image creation, autonomous agents to semantic search - all orchestrated through a unified API. list: - text: OpenAI API compatibility @@ -171,13 +174,40 @@ imageText: - text: Multiple model support icon: hub + - text: Image understanding + icon: image + + - text: Image generation + icon: image + + - text: Audio generation + icon: music_note + + - text: Voice activity detection + icon: mic + + - text: Speech recognition + icon: mic + + - text: Video generation + icon: movie + - text: Privacy focused - icon: lock + icon: security + + - text: Autonomous agents with [LocalAGI](https://github.com/mudler/LocalAGI) + icon: smart_toy + + - text: Semantic search with [LocalRecall](https://github.com/mudler/LocalRecall) + icon: psychology + + - text: Agent orchestration + icon: hub image: path: "images" - filename: "localai_architecture.png" - alt: "LocalAI Architecture" + filename: "imagen.png" + alt: "LocalAI Image generation" imgOrder: desktop: 2 @@ -189,7 +219,7 @@ imageText: # Image compare imageCompare: - enable: true + enable: false weight: 30 template: image compare @@ -204,8 +234,8 @@ imageCompare: addCircleBlur: false, showLabels: true, labelOptions: { - before: 'Input', - after: 'Output', + before: 'Dark', + after: 'Light', onHover: false } } @@ -217,17 +247,17 @@ imageCompare: config: { startingPoint: 50, addCircle: true, - addCircleBlur: false, + addCircleBlur: true, showLabels: true, labelOptions: { before: 'Prompt', after: 'Result', - onHover: false + onHover: true } } imagePath: "images/screenshots" - imageBefore: "image_generation_prompt.webp" - imageAfter: "image_generation_result.webp" + imageBefore: "imagen_before.webp" + imageAfter: "imagen_after.webp" - title: Audio Generation config: { diff --git a/docs/layouts/partials/header.html b/docs/layouts/partials/header.html new file mode 100644 index 00000000..7ba513fc --- /dev/null +++ b/docs/layouts/partials/header.html @@ -0,0 +1,57 @@ + +
+
+ + + + +
+ + + + + {{ with $.Scratch.Get "social_list" }} + + {{ end }} + + + + +
+
+
+ \ No newline at end of file diff --git a/docs/static/android-chrome-192x192.png b/docs/static/android-chrome-192x192.png index f83d7f00..b462084e 100644 Binary files a/docs/static/android-chrome-192x192.png and b/docs/static/android-chrome-192x192.png differ diff --git a/docs/static/android-chrome-512x512.png b/docs/static/android-chrome-512x512.png index c5a3be50..2db366d4 100644 Binary files a/docs/static/android-chrome-512x512.png and b/docs/static/android-chrome-512x512.png differ diff --git a/docs/static/apple-touch-icon.png b/docs/static/apple-touch-icon.png index 9628a2da..b820e718 100644 Binary files a/docs/static/apple-touch-icon.png and b/docs/static/apple-touch-icon.png differ diff --git a/docs/static/favicon-16x16.png b/docs/static/favicon-16x16.png index ee4b2846..c1407a0e 100644 Binary files a/docs/static/favicon-16x16.png and b/docs/static/favicon-16x16.png differ diff --git a/docs/static/favicon-32x32.png b/docs/static/favicon-32x32.png index 9580f203..9762f16b 100644 Binary files a/docs/static/favicon-32x32.png and b/docs/static/favicon-32x32.png differ diff --git a/docs/static/favicon.ico b/docs/static/favicon.ico index 05a5fa9e..6d838e78 100644 Binary files a/docs/static/favicon.ico and b/docs/static/favicon.ico differ diff --git a/docs/static/favicon.svg b/docs/static/favicon.svg new file mode 100644 index 00000000..5e881d4b --- /dev/null +++ b/docs/static/favicon.svg @@ -0,0 +1,171 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + diff --git a/docs/static/site.webmanifest b/docs/static/site.webmanifest new file mode 100644 index 00000000..45dc8a20 --- /dev/null +++ b/docs/static/site.webmanifest @@ -0,0 +1 @@ +{"name":"","short_name":"","icons":[{"src":"/android-chrome-192x192.png","sizes":"192x192","type":"image/png"},{"src":"/android-chrome-512x512.png","sizes":"512x512","type":"image/png"}],"theme_color":"#ffffff","background_color":"#ffffff","display":"standalone"} \ No newline at end of file