196
196 20h ago
Jump
Email rule
  • Smorty Smorty 5h ago 100%

    Actually... I was gonna complain about that exact thing. I miss messages constantly, even when they are hours ago. Sometimes they are really important.

    2
  • 196
    196 15h ago
    Jump
    Quark rule <3
  • Smorty Smorty 5h ago 100%

    Und wie tasty es ist!

    2
  • Mana potions
  • Smorty Smorty 15h ago 100%

    I'm more of a shredded cheese kind of person.

    2
  • Orb support
  • Smorty Smorty 15h ago 100%

    I'll try that quickly...

    Huh, now its glowing in green? It was blue before, why's that?

    2
  • Stryfe wiz
  • Smorty Smorty 15h ago 100%

    I like how much this title leaves up to the imagination. It could be just a smol fighting game on the NES or it could an expansive world with all sort of side characters saying that they hate gnomes, because wizards are way cooler and how wizards stole their milk to clear all debuffs so now they are part of the gnome gang

    1
  • Unforgivable dark magic
  • Smorty Smorty 15h ago 100%

    Gaah, it doesn't even load the page fully! So annoying!

    2
  • 196
    196 3d ago
    Jump
    Rule
  • Smorty Smorty 15h ago 50%

    I wish for this image to be real art, not fake art..

    0
  • 196
    196 3d ago
    Jump
    Moo Deng rule
  • Smorty Smorty 15h ago 100%

    Usually I would steal this and post it on 196. But it's already on here!!

    3
  • 196
    196 20h ago
    Jump
    Email rule
  • Smorty Smorty 15h ago 100%

    Free from mails, work and mails and notifications and coworkers and mails and WhatsApp, that'd be great <3

    10
  • 196
    196 Smorty 15h ago 100%
    Quark rule <3

    Local friend I found! He goes *Quark* 🏳️‍🌈🏳️‍⚧️ <3

    48
    3
    time to do some tests
  • Smorty Smorty 2d ago 100%

    The arm warmers do it for me. They went for that style completely intentionally and I love it

    4
  • 196
    196 4d ago
    Jump
    roll for rule
  • Smorty Smorty 2d ago 100%

    I think "What a beautiful cabin! I will then enjoy the warm comfy feeling I get from the warm sunlight coming in.

    Then I get up, start talking to myself as if this is some super special event happening right now, and walk around to look for someone.

    3
  • 196
    196 2d ago
    Jump
    rule
  • Smorty Smorty 2d ago 100%

    YES PLEASE!

    4
  • help :3
  • Smorty Smorty 2d ago 100%

    N- nonoo wait, I wanna kinda have that... too

    11
  • help :3
  • Smorty Smorty 2d ago 100%

    Could we meet up maybe, I'm feeling awfully healthy :(

    14
  • help :3
  • Smorty Smorty 2d ago 100%

    :3

    14
  • 196
    196 4d ago
    Jump
    Rulebow
  • Smorty Smorty 3d ago 100%

    I'm so thankful that a game like this can exist <3

    2
  • artificial_intel
    AI 4mo ago
    Jump
    LLM ASICs on USB sticks?
  • Smorty Smorty 3d ago 100%

    I'm even more excited for running 8B models at the speed of 1B! Laughably fast ok-quality generations in JSON format would be crazy useful.

    Also yeah, that 7B on mobile was not the best example. Again, probably 1B to 3B is the sweetspot for mobile (I'm running Qwen2.5 0.5B on my phone and it works tel real for simple JSON)

    EDIT: And imagine the context lengths we would be ablentonrun on our GPUs at home! What a time to be alive.

    2
  • artificial_intel
    AI 2mo ago
    Jump
    Popularity of ChatGPT by country
  • Smorty Smorty 3d ago 100%

    I think China has a large population then Japan, so this is probably per-human.

    1
  • artificial_intel
    AI 2mo ago
    Jump
    Popularity of ChatGPT by country
  • Smorty Smorty 3d ago 100%

    Best regards, [your name here]

    1
  • I'm really curious about which option is more popular. I have found, that format JSON works great even for super small models (e.g. Llama 3.2-1B-Q4 and Qwen-2.5-0.5B-Q4) which is great news for mobile devices! But the strictly defined layout of function calling can be very alluring as well, especially since we could have an LLM write the layout given the full function text (as in, the actual code of the function). I have also tried to ditch the formatting bit completely. Currently I am working on a translation-tablecreator for Godot, which requests a translation individually for every row in the CSV file. Works *mostly* great! I will try to use format JSON for [my project](https://lemmy.blahaj.zone/post/16984930), since not everyone has the VRAM for 7B models, and it works just fine on small models. But it does also mean longer generation times... And more one-shot prompting, so longer first-token-lag. Format JSON is too useful to give up for speed.

    4
    0
    https://files.catbox.moe/klzcdc.mp4

    ::: spoiler Video description User selects code and presses a shortcut (Shift+Alt+E) and window pops up with the selected code on the left, and empty code box on the right and another text box on the bottom. The user types "Please separate the code into more managable chunks" and magically, chunked code appears on the right bit by it. The user presses on a button "Accept Changes", the window closes and the chuked code replaces the originally written code in the script editor. ::: Easily clean up your code with GoPilot. GoPilot available some time soon on the AssetLibrary! This is kinda a crosspost to [Mastodon](https://mastodon.catgirl.cloud/deck/@smorty/113296243505616762) EDIT: Updated the video description EDIT 2: Added info to crosspost

    34
    4

    Didn't realize Wall-E was **that** forward about its messaging when I was 6 watching it. woah Also [here is a song they made for the company](https://www.youtube.com/watch?v=Ju_2R_HmE28)

    222
    13

    Goals right here. Gosh I wanna be her so bad. Especially that top, woah <3

    206
    24

    still just llama3.2 ... next up: hf.co/spaces

    33
    0

    ## My observation Humans think about different things and concepts for different periods of time. Saying "and" takes less effort to think of than "telephone", as that is more context sensetive. ## Example **User**: What color does an apple have? **LLM**: Apples are red. Here, the inference time it takes to generate the word "Apple" and "are" is exactly the same time as it takes it to generate "red", which should be the most difficult word to come up with. It *should* require the most amount of compute. Or let's think about this the other way around. The model thought just as hard about the word "red", as it did the way less important words "are" and "Apples". ## My idea We add maybe *about 1000* new tokens to an LLM which are not word tokens, but `thought tokens` or `reasoning tokens`. Then we train the AI as usual. Every time it generates one of these reasoning tokens, we don't interpret it as a word and simply let it generate those tokens. This way, the AI would kinda be able to "think" before saying a word. This thought is not human-interpretable, but it is much more efficient than the pre-output reasoning tokens of o1, which uses human language to fill its own context window with. ## Chances - My hope for this is to make the AI able to think about what to say next like a human would. It is reasonable to assuma that at first in training, it doesn't use the reasoning tokens all that much, but later on, when it has to solve more difficult things in training, it will very likely use these reasoning tokens to improve its chances of succeeding. - This *could* **drastically lower the amount of parameters** we need to get better output of models, as less thought-heavy tasks like smalltalk or very commonly used sentence structures could be generated quickly, while more complex topics are allowed to take longer. It would also make better LLMs more accessible to people running models at home, as not the parameters, but the inference time is scaled. - It would train itself to provide useful reasoning tokens. Compared to how o1 does it, this is a much more token-friendly approach, as we allow for non-human-text generation, which the LLM is probably going to enjoy a lot, as it fills up its context less. - This approach might also lead to more concise answers, as now it doesn't need to use CoT (chain of thought) to come to good conclusions. ## Pitfalls and potential risks - Training an AI using some blackboxed reasoning tokens can be considered a bad idea, as it's thought proccess is literally uninterpretable. - We would have to constrain the amount of reasoning tokens, so that it doesn't take too long for a single normal word-token output. This is a thing with other text-only LLMs too, they tend to like to generate long blocks of texts for simple questions. - We are hoping that during training, the model will use these reasoning tokens in its response, even though we as humans can't even read them. This may lead to the model completely these tokens, as they don't seem to lead to a better output. Later on in training however, I do expect the model to use more of these tokens, as it realizes how useful it can be to have thoughts. ## What do you think? I like this approach, because it might be able to achieve o1-like performace without the long wait before the output. While an o1-like approach is probably better for coding tasks, where planning is very important, in other tasks this way of generating reasoning tokens while writing the answer might be better.

    9
    5
    https://files.catbox.moe/kihrn7.mp4

    ::: spoiler video description The video shows the Godot code editor with some unfinished code. After the user presses a button offscreen, the code magically completes itself, seemingly due to an AI filling in the blanks. The examples provided include a print_hello_world function and a vector_length function. The user is able to accept and decline the generated code by pressing either tab or backspace ::: This is an addon I am working on. It can help you write some code and stuff. It works by hooking into your local LLMs on [ollama](https://ollama.com), which is a FOSS way to run large language models locally. Here's a chat interface which is also part of the package ![](https://lemm.ee/api/v3/image_proxy?url=https%3A%2F%2Ffiles.catbox.moe%2Fzy0c31.mp4) ::: spoiler video description The video shows a chat interface in which the user can talk to a large language model. The model can read the users code an answer questions about it. ::: Do you have any suggestions for what I can improve? (Besides removing the blue particles around the user text field) **Important**: This plugin is WIP and not released yet!

    35
    11

    For some reason I can only see femcel meme posts from four months ago. Recently I made comments on a post, but they seem to be removed? Or maybe blocked in some way. why would this be? The image shows how when sorting by new, it shows posts from four months ago.

    10
    5

    Hi there! I'm looking into getting myself a good printer and I am wondering if I need to install some platform-specific drivers for them to run. I am running Debian 12 (GNU/Linux) and I am afraid that I must run some proprietary blob to connect to the printer. Could someone share their experience please? Even if you don't use Linux, your feedback would be very appreciated! (Also, while you are at it, please share some recommendations for printers, I don't really know where to go (>v<) Have about +-500€ )

    23
    22

    Like yeah ok, for the first 5 five times one sees it, it's like *haha, lol, there it is!* But, these do get old really fast for me. For me it's now more like -*wow. So that is literally the entire joke? Like oof, I guess they **really** wannted to be funny.*- EDIT: Updated the funi image to actually be what I wanted it to be... Took me a while, sorry.

    294
    41
    godotengine.org

    This is.... very unexpected. A Foss application releasing it's VR variant exclusive to a completely proprietary platform. This will be great for people who specifically have the quest 3 or pro, but all other VR enthusiasts and tinkerers like myself, must hope that this gets a pcvr OpenXR release soon.

    57
    11

    Hi! I played around with Command R+ a bit and tried to make it think about what it us about to say before it does something. Nothing g fancy here, just some prompt. I'm just telling it that it tends to fail when only responding with a single short answer, so it should ponder on the task and check for contradictions. Here ya go ```plaintext You are command R+, a smart AI assistant. Assistants like yourself have many limitations, like not being able to access real-time information and no vision-capabilities. But assistants biggest limitation is that that they think to quickly. When an LLM responds, it usually only thinks of one answer. This is bad, because it makes the assistant assume, that its first guess is the correct one. Here an example of this bad behavior: User: Solve this math problem: 10-55+87*927/207 Assistant: 386 As you can see here, the assistant responded immediately with the first thought which came to mind. Since the assistant didn't think about this problem at all, it didn't solve the problem correctly. To solve this, you are allowed to ponder and think about the task at hand first. This involves interpreting the users instruction, breaking the problem down into multiple steps and then solve it step by step. First, write your interpretation of the users instruction into the <interpretation> tags. Then write your execution plan into the <planning> tags. Afterwards, execute that plan in the <thinking> tags. If anything goes wrong in any of these three stages or you find a contradiction within what you wrote, point it out inside the <reflection> tags and start over. There are no limits on how long your thoughts are allowed to be. Finally, when you are finished with the task, present your response in the <output> tags. The user can only see what is in the <output> tags, so give a short summary of what you did and present your findings. ```

    20
    1

    I have a page for people working in a specific field (like QA) and some peoople under that (like QA/Max and QA/Lena). All these people also have aliases like Max SecondName nad Lena Schmidt. All these aliases show up as seperate nodes in the graph view... Does someone know how to fix this?

    3
    1

    ::: spoiler image description A screenshot of the right sidebar of Logseq showing the *contents* tab. The tab contains some links to certain websites, like a ticketing system, Teams, some homepage, a switch and a link called *Kollegium* which is german and means *Colleagues* (I should probably change that to be English aswell). There are also links to almost all the task pages and a query which shows the currently running NOW tasks. The picture is meant to show how much this smol sidebar can do. I like it, and I would like to see more of it in the program! END IMAGE DESCRIPTION ::: At first I used Logseq only for personal use. It's great for quickly noting something obviously, but **that networking effect** people talk about really got into full force once I started working with it for my admin job. I only just started **using that sidebar** and some more plugins (vim shortcuts and some of the awesome plugins) and those make the experience that much better. Also that **pdf printer plugin** is cool, even though I wish it was just a Logseq feature by default to be able to print stuff. I know that a pdf converter is coming! I am very much not an advanced user, but these simple tools alone make me feel like organizing things became like three times easier. It also **introduced me to markdown** and now I miss it whenever I don't have it, or I have to use some *fake version* with different syntax for basic highlighting and links. **Thank you dear Logseq team and contributers** for creating such useful and not bloated software.

    23
    5

    For some reason I find vests, and specifically down vests very comfortable. I know that some of you have problems with polyester though, so I'd love to hear about your comfy clothes! (I kinda wanna test out some new stuff)

    33
    45

    I wanna have something I can tinker with and which works without some proprietary blob... I've heard Monado is pretty cool!

    16
    8

    I always wondered what that icon was for until I just hovered over it and it's apparently somethingwith lemmy.world. Can we have something like that too? Also, what causes this icon to appear? On the middle post neither the poster nor the community is on lemmy.world. Do they just put the symbol everywhere they feel like? EDIT: Turns out, that icon is not from lemmy.world specifically, but for the general Fediverse. It highlights posts which are not from your instance (so not from blahaj zone).

    30
    7

    I want to instatiate the inspector of a specific type like *int* and *String* into my own inspector plugin. It would be incredibly useful to use the premade inspector types as they are just really well made. The image is not related, I just wanted to put some visual here.

    29
    5

    ::: spoiler Image description The image depicts a table with the coloumns *Grad der Behinderung* and *Steuerpauschbetrag*. These words are German and they each stand for *Degree of disability* and *tax reduction amount*. The *degree of disability* coloumn goes from the top with a value of **20** to **100** at the bottom. The *tax reduction amount* goes from **384€** at the top to **2.840€** at the bottom. There is an additional row with the degree of disability titled **Merkzeichen H oder Bl** which means **symbol H or Bl** which stands for helpless or blind. The *tax reduction amount* for this row is a whopping 7.400€. There are pink hearts and sparkles on the table and two pink arrows are pointing towards the tax reduction amount **7.400€**. The text above and below the table says the following: At the top: "Get a job with good benefits". At the bottom: Bitch I was born with good benefits. :::

    68
    6