Skip Navigation
vluz vluz @kbin.social
Posts 4
Comments 29
What're some of the dumbest things you've done to yourself in Linux?
  • Messing around with system python/pip and newly installed versions till all was broken and then looking at documentation.
    This was way back on the 00's and I'm still ashamed on how fast completely I messed it up.

  • WeatherStar 4000+ Emulator
  • Just figured out there are 10 places called Lisbon dotted around the US, according to the search.

  • Linode Alternative Suggestions for Small Projects
  • I got cancelled too and chose Hetzner instead. Will not do business with a company that can't get their filters working decently.

  • Fallout 1 artwork by Justin Sweet
  • Not close enough for V.A.T.S.

  • Pyradion, internet radio TUI client, with recording functionality, written in Python
  • Lovely! I'll go read the code as soon as I have some coffee.

  • Why it’ll be hard to tell if AI ever becomes conscious
  • That is extremely better. It is a very interesting problem, as you put it.

  • Why it’ll be hard to tell if AI ever becomes conscious
  • We know remarkably little about how AI systems work

    Every single time I see this argument used, I stop reading.

  • SSD-1B: 50% smaller and 60% faster Open-Source SDXL Distilled Model
  • I do SDXL generation in 4GB at extreme expense of speed, by using a number of memory optimizations.
    I've done this kind of stuff since SD 1.4, for the fun of it. I like to see how low I can push vram use.

    SDXL takes around 3 to 4 minutes per generation including refiner but it works within constraints.
    Graphics cards used are hilariously bad for the task, a 1050ti with 4GB and a 1060 with 3GB vram.

    Have an implementation running on the 3GB card, inside a podman container, with no ram offloading, 1 vcpu and 4GB ram.
    Graphical UI (streamlit) run on a laptop outside of server to save resources.

    Working on a example implementation of SDXL as we speak and also working on SDXL generation on mobile.
    That is the reason I've looked into this news, SSD-1B might be a good candidate for my dumb experiments.

  • Microsoft Needs So Much Power to Train AI That It's Considering Small Nuclear Reactors
  • Oh my Gwyn, this comment section is just amazing.

  • *Permanently Deleted*
  • Goddammit! Don't tell that one, I use it to impress random people at parties.

  • What’s a game you consider cozy that others might not?
  • Not joking, although I understand it seems very silly at face value.
    Dark Souls 3 PvP specifically SL60+6 at gank town (after pontiff).
    It used to be my go-to wind down after a work day.
    It made me smile and actually relaxed me enough to go to bed and sleep, especially after a hard day.

  • How to set up Podman with NVIDIA GPU acceleration and macvlan networking on Gentoo
  • That's wonderful to know! Thank you again.
    I'll follow your instructions, this implementation is exactly what I was looking for.

  • How to set up Podman with NVIDIA GPU acceleration and macvlan networking on Gentoo
  • Absolutely stellar write up. Thank you!

    I have a couple of questions.
    Imagine I have a powerful consumer gpu card to trow at this solution, 4090ti for the sake of example.
    - How many containers can share one physical card, taking into account total vram memory will not be exceeded?
    - How does one virtual gpu look like in the container? Can I run standard stuff like PyTorch, Tensorflow, and CUDA stuff in general?

  • Recommendations for a context aware text classifier
  • Oof, pop-culture references are hard and I had not considered that at all.
    Thanks for the examples, I'll have a think on how to deal with those.

    My only insight is one you already had.
    Test at least the comment before, and then use the output to dampen or amplify the final result.
    Sorry for being no help at all.

    --

    My project is very basic but I'll post it here for any insight you might get out of it.
    I teach Python in a variety of settings and this is part of a class.

    The data used is from Kaggle: https://www.kaggle.com/competitions/jigsaw-toxic-comment-classification-challenge/
    The original data came from Wikipedia toxic comments dataset.
    There is code too from several users, very helpful for some insight into the problem.

    Data is dirty and needs clean up so I've done so and posted result on HF here:
    https://huggingface.co/datasets/vluz/Tox

    Model is a very basic TensorFlow implementation intended for teaching TF basics.
    https://github.com/vluz/ToxTest
    Some of the helper scripts are very wonky, need fixing before I present this in class.

    Here are my weights after 30 epochs:
    https://huggingface.co/vluz/toxmodel30

    And here is it running on a HF space:
    https://huggingface.co/spaces/vluz/Tox

  • Teaching a free Python course at local charity. Need help with idea.

    Hello!

    I'm teaching Python from scratch at local charity this September. I'll use existing lessons from other Python courses I've done in the past.

    I need your help with the starting project. In my other courses I usually do a mix of text editor, flashcards, note taking, to-do list, calculator, dice rolling, quiz game, etc. Those have worked in the past and will do the job, but are kinda lame. I'll will still go trough some of those as there are lessons there.

    Looking for something that will capture attention at start. Trying to create a "Aha!" moment for the students.

    Is there a perfect starter app that can be built up from a small start?

    Thank you for any answers!

    1
    Recommendations for a context aware text classifier
  • While designing a similar classifier, I've considered the idea of giving it the whole thread as "context" of sorts.
    Not just the parent comment, the whole thread up to original post.

    I've abandoned the idea.
    A comment must stand on it's own, and it would put limits on results, the way I was planning to do it.
    I might be very wrong, your insight into this would be very helpful.

    My original idea was to go recursively trough the thread and test each comment individually.
    Then I would influence the actual comment results with the combined results of it's parents.
    No context during inference, just one comment at a time.

    For example consider thread OP->C1->C2->C3.
    My current model takes milliseconds per test with little resources used.
    It would be ok up to very large threads but would contain a limit to save on answer time.
    I want to determine if Comment 3 is toxic in the context of C2, C1, and OP.
    Test C3, test C2, test C1, test OP. Save results.
    My current model gives answer in several fields ("toxic", "severe toxic", "obscene", "threat", "insult", and "identity hate")
    The idea was to then combine the results of each into a final result for C3.

    How to combine? Haven't figure it out but it would be results manipulation instead of inference/context, etc.

    Edit: Is there any way you can point me at examples difficult to classify? It would be a nice real world test to my stuff.
    Current iteration of model is very new and has not been tested in the wild.

  • Dark fantasy resources

    Hi,

    For a media project, I need to create dark fantasy themed backgrounds. It will be heavily inpainted to meet each background needs.

    I'm looking for models, loras, styles, examples, tutorials, etc. Anything you can think around dark fantasy is valuable, please feel free to suggest linked or related subjects or resources.

    Thanks for help in advance.

    1