Skip Navigation
Fingerthief SimpleDev @infosec.pub

Web Dev Person / Ex Performance ECU Calibrations Person

Posts 8
Comments 44
MinimalChat Is a Full-Featured and Self-Contained LLM Chat Application
  • That seems like a pretty naive and biased approach to software to me honestly.

    Ease of use, community support, feature set, CI/CD etc..all should come into play when deciding what to use.

    Freedom at all costs is great until you limit the community development and potential user base by 90% by using a completely open repo service that 5% of the population uses or some small discord alternative.

    So then the option is to host on multiple platforms/communities and the management and time investment goes up keeping them in sync and active.

    As with most things in life, it's best to look at things with nuance rather than a hard stance imo.

    I may stand it up on another service at some point, but also anyone else is totally free to do that as well. There are no restrictions.

  • MinimalChat Is a Full-Featured and Self-Contained LLM Chat Application
  • I'm not sure I understand at all?

    It's fully open source, can run/connect any number of fully local models as well as the big name models if a user chooses to use them.

    Can you expand on what you mean?

  • MinimalChat Is a Full-Featured and Self-Contained LLM Chat Application
  • Thanks!

    Unfortunately currently there isn't a true RAG implementation largely due to the fact that this site/app is fully self contained with no additional servers or database etc..which is typically required for RAG.

    For now file uploads are stored in the browser's own local database and the content can be extracted and added to the current conversation context easily.

    I definitely want to add a more full RAG system but it's a process to say the least, and if I implement it I want it to be quite effective. My experience with RAG generally has left me quite unimpressed with a few quite decent implementations being the exception.

  • MinimalChat Is a Full-Featured and Self-Contained LLM Chat Application
  • Web search is definitely something I want to add, haven't quite figured out the route I want to take implementing it just yet though.

    Hopefully I can get it added sooner rather than later!

  • MinimalChat Is a Full-Featured and Self-Contained LLM Chat Application

    github.com GitHub - fingerthief/minimal-chat: MinimalChat is a lightweight, open-source chat application that allows you to interact with various large language models.

    MinimalChat is a lightweight, open-source chat application that allows you to interact with various large language models. - fingerthief/minimal-chat

    GitHub - fingerthief/minimal-chat: MinimalChat is a lightweight, open-source chat application that allows you to interact with various large language models.

    cross-posted from: https://infosec.pub/post/13676291

    > I've been building MinimalChat for a while now, and based on the feedback I've received, it's in a pretty decent place for general use. I figured I'd share it here for anyone who might be interested! > > ### Quick Features Overview: > > * Mobile PWA Support: Install the site like a normal app on any device. > * Any OpenAI formatted API support: Works with LM Studio, OpenRouter, etc. > * Local Storage: All data is stored locally in the browser with minimal setup. Just enter a port and go in Docker. > * Experimental Conversational Mode (GPT Models for now) > * Basic File Upload and Storage Support: Files are stored locally in the browser. > * Vision Support with Maintained Context > * Regen/Edit Previous User Messages > * Swap Models Anytime: Maintain conversational context while switching models. > * Set/Save System Prompts: Set the system prompt. Prompts will also be saved to a list so they can be switched between easily. > > The idea is to make it essentially foolproof to deploy or set up while being generally full-featured and aesthetically pleasing. No additional databases or servers are needed, everything is contained and managed inside the web app itself locally. > > It's another chat client in a sea of clients but it is unique in its own ways in my opinion. Enjoy! Feedback is always appreciated! > > Self Hosting Wiki Section https://github.com/fingerthief/minimal-chat/wiki/Self-Hosting-With-Docker >

    I thought sharing here might be a good idea as well, some might find it useful!

    I've added some updates since even the initial post which gave a huge improvement to message rendering speed as well as added a plethora of new models to choose from and load/run fully locally in your browser (Edge and Chrome) with WebGPU and WebLLM

    15
    MinimalChat Is a Full-Featured and Self-Contained LLM Chat Application
  • This project is entirely web based using Vue 3, it doesn't use langchain and I haven't looked into it before honestly but I do see they offer a JS library I could utilize. I'll definitely be looking into that!

    As a result there is no LLM function calling currently and apps like LM Studio don't support function calling when hosting models locally from what I remember. It's definitely on my list to add the ability to retrieve outside data like searching the web and generating a response with the results etc..

  • MinimalChat Is a Full-Featured and Self-Contained LLM Chat Application
  • Yep that's a pretty good comparison!

    I'm curious on what you mean by sourcing training data in an ethical way? I know OpenAI has come under well deserved scrutiny for apparently using content that is hidden behind paywalls without purchasing it themselves in their training data. Which is quite unethical, but aside from that instance I'm interested in hearing some other concerns for my own education.

    In general there are definitely loads of models on places like Hugging Face that are fully open source and provide training data sources for many.

    I believe for Microsoft's new Phi 3 models they actually generated synthetic data themselves for training as well which is an interesting approach that seems to yield good results.

    In the open source LLM world the new Meta Llama 3 models are the latest and greatest, I haven't seen any cause for concerns with it yet. Might be worth looking into those!

  • MinimalChat Is a Full-Featured and Self-Contained LLM Chat Application
  • I haven't personally tried it yet with Ollama but it should work since it looks like Ollama has the ability to use OpenAI Response Formatted API https://github.com/ollama/ollama/blob/main/docs/openai.md

    I might give it go here in a bit to test and confirm.

  • MinimalChat Is a Full-Featured and Self-Contained LLM Chat Application
  • Local models are indeed already supported! In fact any API (local or otherwise) that uses the OpenAI response format (which is the standard) will work.

    So you can use something like LM Studio to host a model locally and connect to it via the local API it spins up.

    If you want to get crazy...fully local browser models are also supported in Chrome and Edge currently. It will download the selected model fully and load it into the WebGPU of your browser and let you chat. It's more experimental and takes actual hardware power since you're fully hosting a model in your browser itself. As seen below.

  • MinimalChat Is a Full-Featured and Self-Contained LLM Chat Application
  • This app is more of an interface to use while connecting to any number of LLM Models that have an API available. The application itself has no model.

    For example you can choose to use GPT-4 Omni by providing an API key from OpenAI.

    But you can also connect to services like OpenRouter with an API key and select between 20+ different models that they provide access to as seen below

    It also supports connecting to fully local models via programs like LM Studio which downloads models from Hugging Face to your machine and will spin up a local API to connect and chat with the model.

  • MinimalChat Is a Full-Featured and Self-Contained LLM Chat Application

    github.com GitHub - fingerthief/minimal-chat: MinimalChat is a lightweight, open-source chat application that allows you to interact with various large language models.

    MinimalChat is a lightweight, open-source chat application that allows you to interact with various large language models. - fingerthief/minimal-chat

    GitHub - fingerthief/minimal-chat: MinimalChat is a lightweight, open-source chat application that allows you to interact with various large language models.

    I've been building MinimalChat for a while now, and based on the feedback I've received, it's in a pretty decent place for general use. I figured I'd share it here for anyone who might be interested!

    Quick Features Overview:

    • Mobile PWA Support: Install the site like a normal app on any device.
    • Any OpenAI formatted API support: Works with LM Studio, OpenRouter, etc.
    • Local Storage: All data is stored locally in the browser with minimal setup. Just enter a port and go in Docker.
    • Experimental Conversational Mode (GPT Models for now)
    • Basic File Upload and Storage Support: Files are stored locally in the browser.
    • Vision Support with Maintained Context
    • Regen/Edit Previous User Messages
    • Swap Models Anytime: Maintain conversational context while switching models.
    • Set/Save System Prompts: Set the system prompt. Prompts will also be saved to a list so they can be switched between easily.

    The idea is to make it essentially foolproof to deploy or set up while being generally full-featured and aesthetically pleasing. No additional databases or servers are needed, everything is contained and managed inside the web app itself locally.

    It's another chat client in a sea of clients but it is unique in its own ways in my opinion. Enjoy! Feedback is always appreciated!

    Self Hosting Wiki Section https://github.com/fingerthief/minimal-chat/wiki/Self-Hosting-With-Docker

    14

    Sending It Around Oulton Park in a Spoon Honda S2000

    For anyone who wants to join, I run a casual AC server for hot lapping competitions.

    Nothing fancy, new track/cars every week or so.

    https://gofast.emperorservers.com/live-timing leaderboards and server join link etc...

    0
    Stowable Simracing Rig: A Foldable Solution for When Space is Tight!
  • I use the F-GT Lite as my main rig and have no real complaints. It's everything I would want from a midrange foldable rig.

    I do want to add a bass shaker or two and I think it'll be a perfectly reasonable cockpit for 95% of people.

    Edit - I of course would always welcome as solid of a wheel mounting system as possible. The Lite has some wiggle but it can be compensated for with few tweaks.

  • Google removes fake Signal and Telegram apps hosted on Play
  • Interesting, thanks for the info!

    I wasn't aware of the update process being used as an attack vector (if it's still a thing) gonna have to read up more on that.

  • Google removes fake Signal and Telegram apps hosted on Play
  • I used Apple for the last few years until recently and I can't say I've ever really noticed stuff like apps faking being another app. That's not to say it doesn't happen of course.

    I do know the Apple app approval process is definitely more strict than what is required for the Play Store.

    I'm not very experienced with Apple or Android development so I'd be curious to hear from devs that use both platforms as well.

  • Trading Paints Credentials Leaked - CHANGE PASSWORDS
  • Wow they really didn't try at all to be even somewhat secure.

  • Hot Lap Challenge | Round 4
  • Well you'll probably really enjoy This Video haha!

    The MGU-H mode was on the "motor" so it doesn't count for this challenge but still that's the absolute most I can get out of the car!

    I'm right on the edge of my ability through a lot of this lap, proper sweaty attempt.

    I'll run in the correct mode for an official lap if I end up needing to, no worries there. At most the advantage was a few tenths as ERS deployment per lap is limited.

    Moar people need to join up and give it a go!

  • Hot Lap Challenge | Round 4
  • Yessir, there's a bit more left on the table but I'm pretty happy with a ~29.8!

    At least until someone beats me hah

  • Version 3.3.0 - Messages UI Tweaks

    Changes from release notes

    • Adjusted chat message bubbles max width to take up nearly the entire width of the chat.

    • Increase sized of message label logos and font.

    • Adjusted message font size and line-height for a better reading experience

    • Added a border to one side of message bubbles for some UI design changes

    0
    Hot Lap Challenge | Round 4
  • Hah, funnily enough Alonso is my all time favorite driver and I have admired his driving style for many years.

    • Yeah Degna 1 is flat if you just close your eyes and huck it in!

    • I don't think I've ever taken Spoon curve in any game or car and thought I did it right, it always feels so wrong to me. I think that attempt was the closest I've been to feeling I got it right.

    The 1:29's are clearly within grasp for both of us, I feel I left a few tenths on the table at the hairpin alone with the lockup and horrible line.

  • Hot Lap Challenge | Round 4
  • Very nice lap! It definitely looks like you tend to be a driver with really smooth and clean inputs, I tend to wrestle things a bit as least in qualifying I've noticed in comparison. I'm always fascinated by driving style differences between drivers while also being relatively equal in speed.

    I just happened to have a set a new lap myself, just to give you a bit of a headache 😆

  • Hot Lap Challenge | Round 4
  • Sounds good to me!

  • Hot Lap Challenge | Round 4
  • You know....I may have ran in hotlap mode as well just out of quali habit.

    I'll have to run again in the default mode this weekend and post that time just to be sure. I don't want any advantage haha

    Good thing you pointed that out lol

    Edit - Though I would argue for hotlaps it would make sense to be able to use whatever default onboard controls the vehicle has imo.

  • Release Notes - Version 3.3.2

    github.com Release Release V3.3.2 - Further Messaging UI Design Improvements · fingerthief/minimal-chat

    Continued improvements to the messaging UI design Updated message bubble padding Updated message bubble min width Updated message bubble border radius to be much less rounded Updated the message la...

    Release Release V3.3.2 - Further Messaging UI Design Improvements · fingerthief/minimal-chat
    0

    MinimalGPT - Features and Guide

    github.com GitHub - fingerthief/minimal-chat: MinimalChat is a lightweight, open-source chat application that allows you to interact with various large language models.

    MinimalChat is a lightweight, open-source chat application that allows you to interact with various large language models. - fingerthief/minimal-chat

    GitHub - fingerthief/minimal-chat: MinimalChat is a lightweight, open-source chat application that allows you to interact with various large language models.

    I've created a fairly thorough overview of MinimalGPT with all the basic info to get started. Please feel free to take a look!

    0

    MinimalGPT Website (PWA)

    Link to a live version of MinimalGPT that I host, you can always spin up a local version youself via the GItHub project.

    0

    He’s Got That Look in His Eyes

    1