

I believe you set env vars on Windows through System Properties -> Advanced -> Environment Variables.
I believe you set env vars on Windows through System Properties -> Advanced -> Environment Variables.
I believe you just need to set the env var OLLAMA_HOST
to 0.0.0.0:11434
and then restart Ollama.
What OS is your server running? Do you have an Android phone or an iPhone?
In either case all you likely need to do is expose the port and then access your server by IP on that port with an appropriate client.
In Ollama you can expose the port to your local network by changing the bind address from 127.0.0.1 to 0.0.0.0
Regarding clients: on iOS you can use Enchanted or Apollo to connect to Ollama.
On Android there are likely comparable apps.
Please, enlighten me - how do you propose we use the term “AI” in a way that’s more useful than a definition that includes machine learning, large language models, and computer vision?
I doubt I’ll agree with your definition, but I’m curious to see how you would exclude machine learning, computer vision, LLMs, etc., from your definition. My assumption is that your definition is going to be either a derivative of “AI is anything computers can’t do yet” or based on pop culture / sci fi, but maybe you’ll surprise me.
To be clear, I’m a software engineer; I’m not speaking in sales speak. I’ve derived my understanding of the term from a combination of its historical context and how it’s used in both professional and academic contexts, not from marketing propaganda or from sci fi and pop culture. I’m certainly aware of the hype machine that’s ongoing, but there are also tons of fascinating advancements happening on a regular basis, and the term “AI” is at minimum a useful term to refer to technologies that leverage similar techniques.
I think of the Bambu P1S as the inexpensive alternative to the Bambu X1C or a comparable printer from Prusa, at least in terms of print consistency and ease of use.
My Bambu was my fourth 3D printer (second FDM printer) and it took 3D printing from a frustrating, time consuming hobby to just a thing I do to enable other hobbies. I don’t have to spend time tweaking settings to get a decent print, because the default settings are already good enough. Instead, I can focus on designing models or working with finished prints.
it’s not ‘ai’, it’s just a poorly trained voice recognition system that’s trying to decipher any random person’s voice.
I’m baffled that you can say “It’s not ‘AI,’ it’s a machine learning powered speech to text system” with a straight face.
Even if we were to agree that ML-powered speech to text isn’t AI (and I don’t agree to that premise, for the record), there’s still the matter of processing the transcription to transform it into something that can be understood by the point of sale system - aka natural language processing. And while that NLP could be implemented without use of an LLM, given LLM’s current level of hype and the ease with which they can be shoved into any given product, I wouldn’t bet on Taco Bell execs approving such an approach, much less asking for it.
I’ll have to check out both OpenSCAD and Code Comic. Some completely non-CAD DSLs that you might be interested in, since you mentioned GraphViz:
Mermaid.js does something very similar to Graphviz. There are a couple other similar tools like that out there, but Mermaid is supported in a lot of places natively or as an easy to use plugin, like GitHub Markdown (and other git forges like Forgejo), Hedgedoc, Obsidian, SilverBullet, etc…
I’d also argue that LaTeX counts, and to a lesser extent, Markdown - compare using them to using Word.
And reveal.js is an equivalent for slide deck creation that would normally be done with PowerPoint.
I hear more complaints about Windows from Windows users than from people who solely or primarily use other OSes. Unless you count “Okay… so why don’t you do something about it?” as a complaint, that is.
I think that makes you “the guy who really likes to talk about Linux.”
From https://wiki.servarr.com/
Welcome to the consolidated wiki for Lidarr, Prowlarr, Radarr, Readarr, Sonarr, and Whisparr. Collectively they are referred to as “*Arr”, “*Arrs”, “Starr”, or “Starrs”. They are designed to automatically grab, sort, organize, and monitor your Music, Movie, E-Book, or TV Show collections for Lidarr, Radarr, Readarr, Sonarr, and Whisparr; and to manage your indexers and keep them in sync with the aforementioned apps for Prowlarr.
See also https://wiki.ravianand.me/home-server/apps/servarr
Servarr is the name for the ecosystem of apps that help you run and automate your own home media server. This includes fetching movie and TV show releases, books and music management, indexer and UseNet/Torrent managers and downloaders.
I could’ve sworn I’d used Foobar2000 on Linux years ago and now I feel like I’m experiencing a mini Mandela effect
Fascinating, thanks for sharing. I didn’t check for every one of those but surprisingly the ones I did check, VLC doesn’t support.
Apparently I should have asked if you’d tried foobar2000, because it has support for all of those, or Audio Overload, which has support for many of them.
PSF
Interesting, it appears Winamp supported PSF via a plugin that basically handled hardware emulation. I found a still open ticket from 2015 for adding support to VLC, though.
According to https://www.vgmpf.com/Wiki/index.php?title=PSF, foobar2000, which has a Linux client, has support. I’ve used foobar2000 before and it’s decent.
Audio Overload is also listed, with a parenthetical - though it’s possible that support has improved since the article was last updated (in 2019). I’ve never used it myself, though.
NSF
Per https://www.vgmpf.com/Wiki/index.php?title=NSF the same players are available, this time without a warning on Audio Overload (notably this article is from 2022). Nosefart is also listed as supporting it and having Linux support.
2SF
https://www.vgmpf.com/Wiki/index.php?title=2SF only lists foobar2000 and Winamp
Various PCM Streams
That’s a lot - and I suspect some of those are supported by VLC based off the codecs listed - but according to https://github.com/vgmstream/vgmstream, foobar2000 has a plugin for vgmstream.
VGM
https://www.vgmpf.com/Wiki/index.php?title=VGM lists foobar2000 and Audio Overload, as well as VGMPlay, which I’ve never heard of before.
GBS
https://www.vgmpf.com/Wiki/index.php?title=GBS again lists foobar2000 and Audio Overload
SPC
https://www.vgmpf.com/Wiki/index.php?title=SPC - same deal.
What kinds of formats does Winamp support that VLC doesn’t support?
Case in point, I have no clue what you wrote, but the intent is clear:
What the fuck did you just fucking say about me, you little bitch? I’ll have you know I graduated top of my class in the Navy Seals, and I’ve been involved in numerous secret raids on Al-Quaeda, and I have over 300 confirmed kills. I am trained in gorilla warfare and I’m the top sniper in the entire US armed forces. You are nothing to me but just another target. I will wipe you the fuck out with precision the likes of which has never been seen before on this Earth, mark my fucking words. You think you can get away with saying that shit to me over the Internet? Think again, fucker. As we speak I am contacting my secret network of spies across the USA and your IP is being traced right now so you better prepare for the storm, maggot. The storm that wipes out the pathetic little thing you call your life. You’re fucking dead, kid. I can be anywhere, anytime, and I can kill you in over seven hundred ways, and that’s just with my bare hands. Not only am I extensively trained in unarmed combat, but I have access to the entire arsenal of the United States Marine Corps and I will use it to its full extent to wipe your miserable ass off the face of the continent, you little shit. If only you could have known what unholy retribution your little “clever” comment was about to bring down upon you, maybe you would have held your fucking tongue. But you couldn’t, you didn’t, and now you’re paying the price, you goddamn idiot. I will shit fury all over you and you will drown in it. You’re fucking dead, kiddo.
Not sure why you’ve gotten downvoted for that, as it’s part of the referenced rule and also true. Unless you’re someone who sees a word in a foreign language and has their brain turn off in response, this should be intelligible to someone who understands English and who doesn’t understand Spanish.
It helps that more than half the words are in English / are used by English speakers: Steam, Proton, Grand Theft Auto 5, Gabe Newell, Linux Mint, Microsoft, Windows, RAM, 100 FPS, 75 FPS
And the important Spanish words are easy to understand:
“Gracias” is pretty commonly understood even by bon-Spanish speakers.
“Uso Software Libre” is pretty obvious, since Libre is a term used in FOSS communities. “Uso” is the most complicated part and I suspect if I didn’t know Spanish I’d just think it meant “Use,” and “Use Libre Software!” is close enough to the intended meaning
Unless Telemetria doesn’t mean Telemetry, it’s pretty obvious.
If I blanked out all the other Spanish words I think the effect would be pretty much the same.
I’m a professional software engineer and I’ve been in the industry since before Kubernetes was first released, and I still found it overwhelming when I had to use it professionally.
I also can’t think of an instance when someone self-hosting would need it. Why did you end up looking into it?
I use Docker Compose for dozens of applications that range in complexity from “just run this service, expose it via my reverse proxy, and add my authentication middleware” to “in this stack, run this service with my custom configuration, a custom service I wrote myself or forked, and another service that I wrote a Dockerfile for; make this service accessible to this other service, but not to the reverse proxy; expose these endpoints to the auth middleware and for these endpoints, allow bypassing of the auth middleware if an API key is supplied.” And I could do much more complicated things with Docker if I needed to, so even for self-hosters with more complex use cases than mine, I question whether Kubernetes is the right fit.
You can store passkeys in (and use them from) a password manager instead of the OS’s secret vault. I think most major password managers support this now - Bitwarden definitely does.
I genuinely don’t understand why people here are taking it so hard that I wish the Immich devs were using semver.
Because you didn’t say that; you said “Breaking changes in a point release? Not cool” and later “I’m basing this off the guidelines at semver.org.”
I’m paraphrasing your comments from memory, to be clear, so apologies if I misquoted you.
It certainly felt to me like you were assuming that this project was using semver and was not following it well, not that you wouldn’t want to use a project that receives this many breaking changes / that doesn’t follow semver. Those complaints both make a lot more sense to me - and I’ve seen many people say similar things about Immich in the past. In fact, it’s a big part of why I haven’t migrated from Photoprism to Immich myself - in this regard they’re complete opposites.
I don’t think there’s any room to argue that announcing a 1.x with a change the developers say is a breaking change, which is what Immich have done, fits within the semver.org guidelines.
That wasn’t the argument.
Following semver is optional. If a project doesn’t explicitly state it is following semver, it shouldn’t be assumed that it is. With regard to Immich in particular, a cursory review of their documentation makes it clear that they are not following semver. Literally, go to https://immich.app/ and read the text at the very top of the page:
⚠️ The project is under very active development. Expect bugs and changes.
Go to the repo and you’ll see the README, which states at the very top:
- ⚠️ The project is under very activedevelopment.
- ⚠️ Expect bugs and breaking changes.
If you can read that, see that they’re on major version 1 with a minor version over 100, and you still think they’re using semver, then that’s on you.
The devs have stated they won’t be using semver until they consider Immich production ready, and that moving to a 1.x version from 0.x was a mistake made some time ago. If you want to think about it as though it is semver, consider the major version to still be 0. See https://github.com/immich-app/immich/discussions/5086#discussioncomment-7593227 for example.
As this project is clearly not following semver, the semver guidelines aren’t applicable and haven’t been violated.
I don’t think there’s any room to argue
Even if semver were applicable, in this case, I would still disagree. The text from semver.org states:
8. Major version X (X.y.z | X > 0) MUST be incremented if any backward incompatible changes are introduced to the public API.
It doesn’t state that any backward incompatible changes, period, require a major version increase, only changes to the public API. I would personally argue that the deployment configuration is part of the public API, but not all project owners agree with me. Even if they do agree, they might say that this was not a documented deployment configuration and thus not part of the public API, and that it therefore doesn’t necessitate an increase to the major version, but as they knew that people were using that configuration, anyway, they included a note about a potentially breaking change as a courtesy to those users.
Can’t you turn on Developer Mode to enable third party slicers?