I wonder if problems could be mostly avoided by running potentially-unsafe code in a container without network access.
Off-and-on trying out an account over at @tal@oleo.cafe due to scraping bots bogging down lemmy.today to the point of near-unusability.
- 31 Posts
- 1.21K Comments
Yes. For a single change. Like having an editor with 2 minute save lag, pushing commit using program running on cassette tapes4 or playing chess over snail-mail. It’s 2026 for Pete’s sake, and we5 won’t tolerate this behavior!
Now of course, in some Perfect World, GitHub could have a local runner with all the bells and whistles. Or maybe something that would allow me to quickly check for progress upon the push6 or even something like a “scratch commit”, i.e. a way that I could testbed different runs without polluting history of both Git and Action runs.
For the love of all that is holy, don’t let GitHub Actions manage your logic. Keep your scripts under your own damn control and just make the Actions call them!
I don’t use GitHub Actions and am not familiar with it, but if you’re using it for continuous integration or build stuff, I’d think that it’s probably a good idea to have that decoupled from GitHub anyway, unless you want to be unable to do development without an Internet connection and access to GitHub.
I mean, I’d wager that someone out there has already built some kind of system to do this for git projects. If you need some kind of isolated, reproducible environment, maybe Podman or similar, and just have some framework to run it?
like macOS builds that would be quite hard to get otherwise
Does Rust not do cross-compilation?
searches
It looks like it can.
https://rust-lang.github.io/rustup/cross-compilation.html
I guess maybe MacOS CI might be a pain to do locally on a non-MacOS machine. You can’t just freely redistribute MacOS.
goes looking
Maybe this?
Darling is a translation layer that lets you run macOS software on Linux
That sounds a lot like Wine
And it is! Wine lets you run Windows software on Linux, and Darling does the same for macOS software.
As long as that’s sufficient, I’d think that you could maybe run MacOS CI in Darling in Podman? Podman can run on Linux, MacOS, Windows, and BSD, and if you can run Darling in Podman, I’d think that you’d be able to run MacOS stuff on whatever.
tal@lemmy.todayto
Technology@lemmy.world•What if the Internet Goes Down? - 15 Jan, 7PM CETEnglish
616·3 days agoI think that Starlink covers a lot of disaster scenarios.
tal@lemmy.todayOPto
Technology@lemmy.world•Zuckerberg eyes massive [datacenter] expansion with Meta Compute playEnglish
8·3 days agoMore AI datacenters being planned.
tal@lemmy.todayto
Technology@lemmy.world•Jensen Huang says relentless negativity around AI is hurting society and has "done a lot of damage"English
22·3 days ago“I think we’ve done a lot of damage with very well-respected people who have painted a doomer narrative, end of the world narrative, science fiction narrative,” Huang said.
I mean, part of the role of science fiction is to try to look at the future as technology changes it. Not all of it is hard sci-fi, but I wouldn’t across-the-board discount sci-fi. A lot of changes have been in sci-fi before reality.
tal@lemmy.todayOPto
Technology@lemmy.world•What are your technology mispredictions?English
4·3 days agoThere are some products out there that do cater to people who want a physical keyboard on their smartphone today. It’s not the norm, but if you’re frustrated over it, it might work for you.
Amazon has a lot of portable Bluetooth keyboards that can basically collapse down into a pocket. Those are generally designed to be used at a table, though, not in a Blackberry-style thumb keyboard sense. I’m pretty sure that I’ve seen a few of the latter that can clip to a phone, though.
tal@lemmy.todayOPto
Technology@lemmy.world•What are your technology mispredictions?English
2·3 days agowhile visiting family in LA I saw a robot waiter which both (optionally) took orders and would serve as a mobile food/plate tray.
Yeah, I’ve run into these. That being said, I’d call them currently a novelty…but I also remember when using touchscreen kiosks for ordering instead of cashiers was a rarity.
tal@lemmy.todayOPto
Technology@lemmy.world•What are your technology mispredictions?English
1·3 days agoThere was some similar project that the UK was going to do, run an HVDC submarine line down from the UK to Africa.
searches
https://en.wikipedia.org/wiki/Xlinks_Morocco–UK_Power_Project
The Xlinks Morocco-UK Power Project is a proposal to create 11.5 GW of renewable generation, 22.5 GWh of battery storage and a 3.6 GW high-voltage direct current interconnector to carry solar and wind-generated electricity from Morocco to the United Kingdom.[1][2][3][4] Morocco has been hailed as a potential key power generator for Europe as the continent looks to reduce reliance on fossil fuels.[5]
If built, the 4,000 km (2,500 miles) cable would be the world’s longest undersea power cable, and would supply up to 8% of the UK’s electricity consumption.[6][7][8] The project was projected to be operational within a decade.[9][10] The proposal was rejected by the UK government in June 2025.
tal@lemmy.todayOPto
Technology@lemmy.world•What are your technology mispredictions?English
1·5 days agoWell, if vinyl could come back, I suppose MiniDiscs can too.
tal@lemmy.todayOPto
Technology@lemmy.world•What are your technology mispredictions?English
5·5 days agoI think another major factor for Linux gaming beyond Valve was a large shift by game developers to using widely-used game engines. A lot of the platform portability work happened at that level, so was spread across many games. Writing games that could run on both personal computers and personal-computer-like consoles with less porting work became a goal. And today, some games also have releases on mobile platforms.
When I started using Linux in the late 1990s, the situation was wildly different on that front.
tal@lemmy.todayOPto
Technology@lemmy.world•What are your technology mispredictions?English
4·5 days agoContext:
https://en.wikipedia.org/wiki/Ultra-mobile_PC
An ultra-mobile PC,[1] or ultra-mobile personal computer (UMPC), is a miniature version of a pen computer, a class of laptop whose specifications were launched by Microsoft and Intel in Spring 2006. Sony had already made a first attempt in this direction in 2004 with its Vaio U series, which was only sold in Asia. UMPCs are generally smaller than subnotebooks, have a TFT display measuring (diagonally) about 12.7 to 17.8 centimetres (5.0 to 7.0 in), are operated like tablet PCs using a touchscreen or a stylus, and can also have a physical keyboard.
tal@lemmy.todayOPto
Technology@lemmy.world•What are your technology mispredictions?English
4·5 days agoconsiders
I’ve been in a couple conversation threads about this topic before on here. I’m more optimistic.
I think that the Internet has definitely democratized information in many ways. I mean, if you have an Internet connection, you have access to a huge amount of information. Your voice has an enormous potential reach. A lot of stuff where one would have had to buy expensive reference works or spend a lot of time digging information up are now readily available to anyone with Internet access.
I think that the big issue wasn’t that people became less critical, but that one stopped having experts filter what one saw. In, say, 1996, most of what I read had passed through the hands of some sort of professional or professionals specialized in writing. For newspapers or magazines, maybe it was a journalist and their editor. For books, an author and their editor and maybe a typesetter.
Like, in 1996, I mostly didn’t get to actually see the writing of Average Joe. In 2026, I do, and Average Joe plays a larger role in directly setting the conversation. That is democratization. Average Joe of 2026 didn’t, maybe, become a better journalist than the professional journalist of 1996. But…I think that it’s very plausible that he’s a better journalist than Average Joe of 1996.
Would it have been reasonable to expect Average Joe of 2026 to, in addition to all the other things he does, also be better at journalism than a journalist of 1996? That seems like a high bar to set.
And we’re also living in a very immature environment as our current media goes. I am not sold that this is the end game.
There’s a quote from Future Shock — written in 1970, but I think that we can steal the general idea for today:
It has been observed, for example, that if the last 50,000 years of man’s existence were divided into lifetimes of approximately sixty-two years each, there have been about 800 such lifetimes. Of these 800, fully 650 were spent in caves.
Only during the last seventy lifetimes has it been possible to communicate effectively from one lifetime to another—as writing made it possible to do. Only during the last six lifetimes did masses of men ever see a printed word. Only during the last four has it been possible to measure time with any precision. Only in the last two has anyone anywhere used an electric motor. And the overwhelming majority of all the material goods we use in daily life today have been developed within the present, the 800th, lifetime.
That’s just to drive home how extremely rapidly the environment in which we all live has shifted compared to how it had in the past. In that quote, Alvin Toffler was talking about how incredibly quickly things had changed in that it had only been six lifetimes since the public as a whole had seen printed text, how much things had changed. But in 2026, we live in a world where it has only been a quarter of a lifetime, less for most, since much of the global population of humanity has been intimately linked by near-instant, inexpensive, mass communication.
I think that it would be awfully unexpected and surprising if we would have immediately figured out conventions and social structures and technical solutions to every deficiency for such a new environment. Social media is a very new thing in the human experience at this scale. I think that it is very probable that humanity will — partly by trial-and-error, getting some scrapes and bruises along the way — develop practices to smooth over rough spots and address problems.
Consider, say, the early motorcar, which had no seatbelts, windscreen, roof, suspension, was driven on a road infrastructure designed for horse-drawn carts to travel maybe ten miles an hour, didn’t have a muffler, didn’t have an electric starter, lacked electric headlights and other lighting, an instrument panel, and all that. It probably had a lot of very glaring problems as a form of transportation to people who saw it. An awful lot of those problems have been solved over time. I think that it would be very surprising if electronic mass communication available to everyone doesn’t do something similar.
tal@lemmy.todayOPto
Technology@lemmy.world•What are your technology mispredictions?English
7·5 days agoI don’t know if I can count this as mine, but I certainly didn’t disagree with predictions of others around 1990 or so that the smart home would be the future. The idea was that you’d have a central home computer and it would interface with all sorts of other systems and basically control the house.
While there are various systems for home automation, things like Home Assistant or OpenHAB, and some people use them, and I’ve used some technology that were expected to be part of this myself, like X10 for device control over power circuits, the vision of a heavily-automated, centrally-controlled home never made it to become the normal. I think that the most-widely-deployed piece of home automation that has shown up since then is maybe the smart thermostat, which isn’t generally hooked into some central home computer.
tal@lemmy.todayOPto
Technology@lemmy.world•What are your technology mispredictions?English
211·5 days ago
tal@lemmy.todayto
Technology@lemmy.world•Dell says the quiet part out loud: Consumers don't actually care about AI PCs — "AI probably confuses them more than it helps them"English
7·7 days agoSo, it’s not really a problem I’ve run into, but I’ve met a lot of people who have difficulty on Windows understanding where they’ve saved something, but do remember that they’ve worked on or looked at it at some point in the past.
My own suspicion is that part of this problem stems from the fact that back in the day, DOS had a not-incredibly-aimed-at-non-technical-users filesystem layout, and Windows tried to avoid this by hiding that and stacking an increasingly number of “virtual” interfaces on top of things that didn’t just show one the filesystem, whether it be the Start menu or Windows Explorer and file dialogs having a variety of things other than just the filesystem to navigate around. The result is that you have had Microsoft banging away for much of the lifetime of Windows trying to add more ways to access files, most of which increase the difficulty of actually understanding what is going on fully through the extra layers. But regardless of why, some users do have trouble with it.
So if you can just provide a search that can summon up that document where they were working on that had a picture of giraffes by typing “giraffe” into some search field, maybe that’ll do it.
tal@lemmy.todayto
Technology@lemmy.world•Bose open-sources its SoundTouch home theater smart speakers ahead of end-of-lifeEnglish
6·7 days agoI don’t use OpenHAB or Home Assistant, but I’d be extremely surprised if they don’t have existing functionality for connecting microphones, speakers, and LLMs to set up voice-controlled stuff.
searches
Willow Is a Practical, Open Source, Privacy-focused Platform for Voice Assistants and Other Applications
Willow is an ESP IDF based project primarily targeting the ESP32-S3-BOX hardware family from Espressif. Our goal is to provide Amazon Echo/Google Home competitive performance, accuracy, cost and functionality with Home Assistant, openHAB and other platforms.
100% open source and completely self-hosted by the user with “ready for the kitchen counter” low cost commercially available hardware.
https://rhasspy.readthedocs.io/en/latest/
Rhasspy (ɹˈæspi) is an open source, fully offline set of voice assistant services for many human languages that works well with:
tal@lemmy.todayto
Technology@lemmy.world•Dell says the quiet part out loud: Consumers don't actually care about AI PCs — "AI probably confuses them more than it helps them"English
11·8 days agoMy understanding from a very brief skim of what Microsoft was doing with Copilot is to take screenshots constantly, run image recognition on it, and then make it searchable as text and have the ability to go back and view those screenshots in a timeline. Basically, adding more search without requiring application-level support.
They may also have other things that they want to do, but that was at least one.
EDIT: They specifically called that feature “Recall”, and it was apparently the “flagship” feature of Copilot.
tal@lemmy.todayto
Technology@lemmy.world•Dell says the quiet part out loud: Consumers don't actually care about AI PCs — "AI probably confuses them more than it helps them"English
49·8 days agoNot the position Dell is taking, but I’ve been skeptical that building AI hardware directly into specifically laptops is a great idea unless people have a very concrete goal, like text-to-speech, and existing models to run on it, probably specialized ones. This is not to diminish AI compute elsewhere.
Several reasons.
-
Models for many useful things have been getting larger, and you have a bounded amount of memory in those laptops, which, at the moment, generally can’t be upgraded (though maybe CAMM2 will improve the situation, move back away from soldered memory). Historically, most users did not upgrade memory in their laptop, even if they could. Just throwing the compute hardware there in the expectation that models will come is a bet on the size of the models that people might want to use not getting a whole lot larger. This is especially true for the next year or two, since we expect high memory prices, and people probably being priced out of sticking very large amounts of memory in laptops.
-
Heat and power. The laptop form factor exists to be portable. They are not great at dissipating heat, and unless they’re plugged into wall power, they have sharp constraints on how much power they can usefully use.
-
The parallel compute field is rapidly evolving. People are probably not going to throw out and replace their laptops on a regular basis to keep up with AI stuff (much as laptop vendors might be enthusiastic about this).
I think that a more-likely outcome, if people want local, generalized AI stuff on laptops, is that someone sells an eGPU-like box that plugs into power and into a USB port or via some wireless protocol to the laptop, and the laptop uses it as an AI accelerator. That box can be replaced or upgraded independently of the laptop itself.
When I do generative AI stuff on my laptop, for the applications I use, the bandwidth that I need to the compute box is very low, and latency requirements are very relaxed. I presently remotely use a Framework Desktop as a compute box, and can happily generate images or text or whatever over the cell network without problems. If I really wanted disconnected operation, I’d haul the box along with me.
EDIT: I’d also add that all of this is also true for smartphones, which have the same constraints, and harder limitations on heat, power, and space. You can hook one up to an AI accelerator box via wired or wireless link if you want local compute, but it’s going to be much more difficult to deal with the limitations inherent to the phone form factor and do a lot of compute on the phone itself.
EDIT2: If you use a high-bandwidth link to such a local, external box, bonus: you also potentially get substantially-increased and upgradeable graphical capabilities on the laptop or smartphone if you can use such a box as an eGPU, something where having low-latency compute available is actually quite useful.
-
tal@lemmy.todayto
Technology@lemmy.world•Samsung, SK Reportedly Hike Server DRAM Prices 60-70% – Google, Microsoft in the QueueEnglish
5·8 days agoI know open ai bought ~40% of microns memory production.
IIRC Micron was the only Big Three DRAM manufacturer that OpenAI didn’t sign a contract with. I think that they signed contracts with SK Hynix and Samsung for their supply, and didn’t with Micron.
searches
Yeah:
OpenAI ropes in Samsung, SK Hynix to source memory chips for Stargate
Not signing was actually probably to Micron’s advantage; I understand that OpenAI didn’t let Samsung know that they were negotiating with SK Hynix and didn’t let SK Hynix that they were negotiating with Samsung and signed both deals concurrently. That is, each of Samsung and SK Hynix probably sold the DRAM that went to OpenAI for less than they could have gotten on the open market, since neither was aware at the time of signing that the supply on the open market outside of themselves would sharply decrease during the period of the contract, which would be expected to drive up prices.
I mean, they still made a lot more money than they had been making. Just that they could have probably managed to get even more money for the DRAM that they sold.
IIRC the 40% number was OpenAI signing for 40% of global production output, not for any particular company’s output.



















I actually don’t know what the current requirement is. Back in the day, Apple used to build some of the OS — like QuickDraw — into the ROMs, so unless you had a physical Mac, not just a purchased copy of MacOS, you couldn’t legally run MacOS, since the ROM contents were copyrighted, and doing so would require infringing on the ROM copyright. Apple obviously doesn’t care about this most of the time, but I imagine that if it becomes institutionalized at places that make real money, they might.
But I don’t know if that’s still the case today. I’m vaguely recalling that there was some period where part of Apple’s EULA for MacOS prohibited running MacOS on non-Apple hardware, which would have been a different method of trying to tie it to the hardware.
searches
This is from 2019, and it sounds like at that point, Apple was leveraging the EULAs.
https://discussions.apple.com/thread/250646417?sortBy=rank
They switched to ARM in 2020, so unless their legal position changed around ARM, I’d guess that they’re probably still relying on the EULA restrictions. That being said, EULAs have also been thrown out for various reasons, so…shrugs
goes looking for the actual license text.
Yeah, this is Tahoe’s EULA, the most-recent release:
https://www.apple.com/legal/sla/docs/macOSTahoe.pdf
Page 2 (of 895 pages):
They allow only on Apple-branded hardware for individual purchases unless you buy from the Mac Store. For Mac Store purchases, they allow up to two virtual instances of MacOS to be executed on Apple-branded hardware that is also running the OS, and only under certain conditions (like for software development). And for volume purchase contracts, they say that the terms are whatever the purchaser negotiated. I’m assuming that there’s no chance that Apple is going to grant some “go use it as much as you want whenever you want to do CI tests or builds for open-source projects targeting MacOS” license.
So for the general case, the EULA prohibits you from running MacOS wherever on non-Apple hardware.