• 0 Posts
  • 109 Comments
Joined 11 months ago
cake
Cake day: May 22nd, 2025

help-circle
  • As a male with an unfortunate genetic condition I can honestly say it’s not just the women.

    Sitting on the sidelines, and watching my friend group jump into bed and into marriage with the first consenting partner is just about the dumbest thing imaginable. Years later they wake up and realize that the hormones experienced during the young adult years, did not allow them to make correct life decisions. But, now there’s a mortgage and kids and they’re stuck and miserable.

    Some cheat, some take out their anger on what they see as the cause of their unhappiness: the wife, the kids. It’s a stupid situation, and the amount of times I’ve heard “Why didn’t you stop me?”… I tried, ad nauseam.





  • … What artificial facade? Canada has a history as a country of brutal soldiers. For us the Geneva Convention is more of a checklist. We’re very polite: Surrendering Germans got a bullet to the head as a thank you.

    We had such a bad reputation the government shut down the Canadian Airborne Regiment (Our equivalent of the Marines) and pivoted us into a peacekeeping role to change perceptions.

    We have upcoming trade renegotiations, and pissing off the orange turd before they start just isn’t in the cards. We’ll support international efforts in a way that doesn’t piss off our brother to the south.



  • Canada and the US aren’t just allies, we’re tightly integrated militarily.

    It may look like two separate countries on the political stage, but behind the scenes we operate together. When Canadian personnel are already embedded in command structures in the region, it’s not realistic to expect Canada to fully denounce the conflict




    1. Fixed credential-exfiltration risk in /api/proxy/image: Previously the endpoint could:
    • accept arbitrary auth_id
    • load stored API keys
    • forward them to attacker-controlled URLs
    1. Enforced outbound host allowlist globally Previously:
    • allowlist existed
    • but outgoingFetch() didn’t enforce it
    • plugins/engines could bypass it
    1. Fixed extension store path traversal Previously a malicious store manifest could:
    • inject … paths
    • escape install directories
    • reference arbitrary files
    1. Hardened proxy IP trust Previously:
    • rate limiting trusted any X-Forwarded-For header
    • clients could spoof their IP
    1. Fixed inconsistent settings authentication Previously:
    • settings UI stored an auth token
    • but the settings modal didn’t send it when saving
    1. Implemented Improved proxy deployment support
    • Added proxy-aware behavior:
    • DEGOOG_PUBLIC_BASE_URL for canonical URLs
    • secure cookie handling when X-Forwarded-Proto=https

    Additional Improvements:

    • suggestion fetching hardened
    • DuckDuckGo suggestion parsing fixed
    • unified outbound request handling
    • install state guard properly cleaned up

    Made some other changes for my specific deployment. Very happy with your work so far. Thanks so much





  • I have kleinfelter syndrome, which means my body does not produce Testosterone, or very little anyway. I only found out at 38. I’ve had a pretty advanced internal monologue all my life, what the lack of T did give me was insulin resistance, major depression, and generalized anxiety. The latter 2 disappeared almost instantly after my first Testosterone injection. Depression was like a constant weight pushing down on me, the next morning I woke up without the weight, and without the worry about everything. And it was very weird.




  • It’s very content specific, what are you processing with the API?

    One of my little side projects right now is translating Russian fiction, specifically a genre over there called ‘boyar-anime’ which is essentially fantasy set in imperial russia. I do most my heavy translation using Anthropic Haiku which is very cheap and unlike the higher end models it tends to dumb down some of the more complex parts of Imperial Russian aristocracy so it’s more in line with similar fiction over here. When I take the source book, I chunk it down into small segments that I translate individually so I don’t get context bleed, then I mechanically process to find anything that didn’t translate very well. I combine roughly 40 of these weirdly translated segments into a jsonl file and submit the file through the API. OpenAI Batch API can accept up to 900k tokens, but you’ll wait close to 11 hours for something that large. 40 segments is around 30k tokens and that usually processes in a few mins to an hour depending.

    The jsonl file is essentially made up of smaller json blocks

    {
      "custom_id": "SEGMENT-NUM",
      "method": "POST",
      "url": "/v1/responses",
      "body": {
        "model": "gpt-5.3",
        "input": [
          {
            "role": "system",
            "content": [
              {
                "type": "input_text",
                "text": "You are a meticulous English language proofreader."
              }
            ]
          },
          {
            "role": "user",
            "content": [
              {
                "type": "input_text",
                "text": "PROMPT - SUBMITTED SEGMENT"
              }
            ]
          }
        ],
        "max_output_tokens": 8192
      }
    }
    

    I then setup polling to check back with the API every few mins, when the submitted queries are completed, I send more automatically until everything has been processed.



  • I don’t know what thinking profile your friend was using but asking ChatGPT that with the mixed tasks profile showed an almost immediate result with absolutely no thinking required.

    LLM’s are a tool, like with any tool there is a learning curve, and in my opinion the majority of “AI” users are unable to use the tool properly, and then get mad at the tool. Or like you, want to disparage the use of an LLM so they bait the LLM with tasks that it knows will fail or hallucinate on. To me that’s like blaming the table saw because it cut off your finger. Do the majority of people need a paid account? No.

    Are there people working in the Tech sector who use an LLM everyday, who have corporate accounts and paid accounts at home for their own projects: absolutely. I know a large number of them, most are Lemmy users as well. But because there is so much negativity from the open source crowd, all these engineers are afraid to discuss all the ways it makes our lives easier. So we get a disproportionate amount of negativity. I’m getting to a point where the amount of AI shit posting on here is like the amount of vegan shit posting on Reddit. And just as stupid.


  • Opus is heavily throttled outside enterprise tiers. I was regularly blowing through weekly usage limits by Tuesday using Opus. 5.3 on the higher thinking profiles match or exceed Opus capabilities, and I have yet to hit a single limitation.

    If I need to process via API I will run tests against Anthropic Haiku or Sonnet before trying Gpt5-mini, If I need to use 5.3, and what I’m doing isn’t time critical I’ll use batch processing. Smaller token batches complete very quickly, often in under 2 hours. And at a 50% discount, provides serious cost savings.