Tootfinder

Opt-in global Mastodon full text search. Join the index!

@tiotasram@kolektiva.social
2025-07-22 00:03:45

Overly academic/distanced ethical discussions
Had a weird interaction with @/brainwane@social.coop just now. I misinterpreted one of their posts quoting someone else and I think the combination of that plus an interaction pattern where I'd assume their stance on something and respond critically to that ended up with me getting blocked. I don't have hard feelings exactly, and this post is only partly about this particular person, but I noticed something interesting by the end of the conversation that had been bothering me. They repeatedly criticized me for assuming what their position was, but never actually stated their position. They didn't say: "I'm bothered you assumed my position was X, it's actually Y." They just said "I'm bothered you assumed my position was X, please don't assume my position!" I get that it's annoying to have people respond to a straw man version of your argument, but when I in response asked some direct questions about what their position was, they gave some non-answers and then blocked me. It's entirely possible it's a coincidence, and they just happened to run out of patience on that iteration, but it makes me take their critique of my interactions a bit less seriously. I suspect that they just didn't want to hear what I was saying, while at the same time they wanted to feel as if they were someone who values public critique and open discussion of tricky issues (if anyone reading this post also followed our interaction and has a different opinion of my behavior, I'd be glad to hear it; it's possible In effectively being an asshole here and it would be useful to hear that if so).
In any case, the fact that at the end of the entire discussion, I'm realizing I still don't actually know their position on whether they think the AI use case in question is worthwhile feels odd. They praised the system on several occasions, albeit noting some drawbacks while doing so. They said that the system was possibly changing their anti-AI stance, but then got mad at me for assuming this meant that they thought this use-case was justified. Maybe they just haven't made up their mind yet but didn't want to say that?
Interestingly, in one of their own blog posts that got linked in the discussion, they discuss a different AI system, and despite listing a bunch of concrete harms, conclude that it's okay to use it. That's fine; I don't think *every* use of AI is wrong on balance, but what bothered me was that their post dismissed a number of real ethical issues by saying essentially "I haven't seen calls for a boycott over this issue, so it's not a reason to stop use." That's an extremely socially conformist version of ethics that doesn't sit well with me. The discussion also ended up linking this post: chelseatroy.com/2024/08/28/doe which bothered me in a related way. In it, Troy describes classroom teaching techniques for introducing and helping students explore the ethics of AI, and they seem mostly great. They avoid prescribing any particular correct stance, which is important when teaching given the power relationship, and they help students understand the limitations of their perspectives regarding global impacts, which is great. But the overall conclusion of the post is that "nobody is qualified to really judge global impacts, so we should focus on ways to improve outcomes instead of trying to judge them." This bothers me because we actually do have a responsibility to make decisive ethical judgments despite limitations of our perspectives. If we never commit to any ethical judgment against a technology because we think our perspective is too limited to know the true impacts (which I'll concede it invariably is) then we'll have to accept every technology without objection, limiting ourselves to trying to improve their impacts without opposing them. Given who currently controls most of the resources that go into exploration for new technologies, this stance is too permissive. Perhaps if our objection to a technology was absolute and instantly effective, I'd buy the argument that objecting without a deep global view of the long-term risks is dangerous. As things stand, I think that objecting to the development/use of certain technologies in certain contexts is necessary, and although there's a lot of uncertainly, I expect strongly enough that the overall outcomes of objection will be positive that I think it's a good thing to do.
The deeper point here I guess is that this kind of "things are too complicated, let's have a nuanced discussion where we don't come to any conclusions because we see a lot of unknowns along with definite harms" really bothers me.

@Zer0Rank@social.linux.pizza
2025-07-23 04:40:06

My golang knowledge is pretty basic. I had to do some updates to a product at work, and it was pretty easy and straight forward. Actually quite fun. Just to learn more golang, I started working on a cli-tool for my Niri setup, and it's actually pretty usable already.
I just want to polish it and clean up the code a bit more before publishing it. Also, since I'm not that good at the language yet, I'm totally not sure if I even do things correctly or the 'golang' way.…

@cheryanne@aus.social
2025-09-22 10:17:45

By the way, what time is the Rapture occurring tomorrow? I have a few things I want to do beforehand, so hopefully it will be later in the day.
#Rapture #Sarcasm

@mgorny@social.treehouse.systems
2025-07-14 16:39:18

About morbid thriftiness (Autism Spectrum Condition)
As you may have noticed, I am morbidly thrifty. Usually I don't buy stuff that I don't need — and if I decide that I actually need something, I am going to ponder about it for a while, look for value products, and for the best price. And with some luck, I'm going to decide I don't need it that bad after all.
One reason for that is probably how I was raised. My parents taught me to be thrifty, so I have to be. It doesn't matter that, from retrospective, I see that their thriftiness was applied rather arbitrarily to some spendings and not others, or that perhaps they were greedy — spending less on individual things so that they could buy more. Well, I can't delude myself like that, so I have to be thrifty for real. And when I fail, when I pay too much, when I get cheated — I feel quite bad about it.
The other reason is that I keep worrying about my future. It doesn't matter how rich I may end up — I'll keep worrying that I'll run out of money in the future. Perhaps I'll lose a job and won't be able to find anything for a long time, Perhaps something terrible will happen and I'm going to need to pay a lot suddenly.
Another thing is that I easily get attached to objects. Well, it's easier to be thrifty when you really don't want to replace stuff. Over time you also learn to avoid getting new stuff at all, since the more stuff you have, the more stuff may break and need to be thrown away.
Finally, there's my environmental responsibility. I admit that I don't do enough — but at least the things I can do, I do.
[EDIT: and yes, I feel bad about how expensive my new phone was, even though it's of much higher quality than the last one. Also, I got a worse deal because I waited too long.]
#ActuallyAutistic

@chris@mstdn.chrisalemany.ca
2025-08-15 16:40:43

2/2 I continued blogging Alberniweather and on FB and Twitter but I gradually removed my personal self from Facebook and eventually during the Pandemic, I decided the Facebook environment was just too toxic even for weather stuff and I shut down my page and left Facebook completely.
The impact on traffic to Alberniweather.ca and its prominence in the community was, and still is, significant.
I have diehard followers, many who have become friends over the years, I still get the odd call from media, or even the public about random weather things.
I have good connections with a few folks at Environment Canada (though their staff have become thinner and more transient :(
and major events still get spikes of local traffic but I since about 2022, and after I removed myself from Twitter that year, I don’t blog nearly as much. I would do a few posts in a week, and then go months without posting. I just got out of the habit I guess.
But I am still interested in the weather. I still feel like Alberniweather is a useful service for people in my community. I still feel a willing obligation to inform people about the weather and I believe I am trusted to do so by the public and local leaders. I’ve never made any money at it, I sold ad space on the website for a few years but it wasn’t worth the hassle and I didn’t feel comfortable taking the money when I was councillor. I have had some generous spontaneous donations at times.
But mainly I do it because it’s interesting, and I hope it is useful for people especially when people are looking for information during a major event.
The highest traffic I have ever had on Alberniweather pre-FB exit was the local Dog Mountain forest fire in 2015.
post-FB exit: the #underwoodfire
People want easy access to reliable local, trusted, information.
Large media orgs have mostly given up on this.
I am grateful we still have an active local newspaper and radio and that both trust me and I trust them.
@… @…

@hex@kolektiva.social
2025-09-11 19:43:39

As a survivor of the type of stochastic terrorism that people like Charlie Kirk carry out, and also a survivor of gun violence, I've had a bit to think about.
I still understand the anger, the frustration, that leads someone to snap. I'll always have compassion, if also paired with some frustration, for the person who shot me and her husband. That makes sense to me. I refused to testify because it never made sense to punish a tool who acted out of ignorance.
What doesn't make as much sense to me are the upper class grifters, the stochastic terrorists, who turn a profit of off setting people like that up to kill. I always wanted to see people like that, people who were clearly profiting off of bringing evil into the world, held accountable.
But the way I always want to see them held accountable, is by being forced to live in a world where everyone is free. Charlie Kirk got out easy, and that is a bit sad. He should have had to suffer through our victory. He should have had to watch the fall of Western Civilization, the collapse of all the things he held dear, all the things he tried to uphold. I wish that he had lived long enough to understand the suffering he inflicted on the world. He should have lived long enough to have to do real work, to have to figure out how to feed himself, house himself, in a world that has no market for the hate that he brings.
As much as I had rather that he starved, I would never shed a tear for the opening of a new public urinal.
Rest in piss. Charlie Kirk.

@tiotasram@kolektiva.social
2025-07-28 13:06:20

How popular media gets love wrong
Now a bit of background about why I have this "engineered" model of love:
First, I'm a white straight cis man. I've got a few traits that might work against my relationship chances (e.g., neurodivergence; I generally fit pretty well into the "weird geek" stereotype), but as I was recently reminded, it's possible my experience derives more from luck than other factors, and since things are tilted more in my favor than most people on the planet, my advice could be worse than useless if it leads people towards strategies that would only have worked for someone like me. I don't *think* that's the case, but it's worth mentioning explicitly.
When I first started dating my now-wife, we were both in graduate school. I was 26, and had exactly zero dating/romantic experience though that point in my life. In other words, a pretty stereotypical "incel" although I definitely didn't subscribe to incel ideology at all. I felt lonely, and vaguely wanted a romantic relationship (I'm neither aromantic nor asexual), but had never felt socially comfortable enough to pursue one before. I don't drink and dislike most social gatherings like parties or bars; I mostly hung around the fringes of the few college parties I attended, and although I had a reasonable college social life in terms of friends, I didn't really do anything to pursue romance, feeling too awkward to know where to start. I had the beginnings of crushes in both high school and college, but never developed a really strong crush, probably correlated with not putting myself in many social situations outside of close all-male friend gatherings. I never felt remotely comfortable enough to act on any of the proto-crushes I did have. I did watch porn and masturbate, so one motivation for pursuing a relationship was physical intimacy, but loneliness was as much of a motivating factor, and of course the social pressure to date was a factor too, even though I'm quite contrarian.
When I first started dating my now-wife, we were both in graduate school. I was 26, and had exactly zero dating/romantic experience though that point in my life. In other words, a pretty stereotypical "incel" although I definitely didn't subscribe to incel ideology at all. I felt lonely, and vaguely wanted a romantic relationship (I'm neither aromantic nor asexual), but had never felt socially comfortable enough to pursue one before. I don't drink and dislike most social gatherings like parties or bars; I mostly hung around the fringes of the few college parties I attended, and although I had a reasonable college social life in terms of friends, I didn't really do anything to pursue romance, feeling too awkward to know where to start. I had the beginnings of crushes in both high school and college, but never developed a really strong crush, probably correlated with not putting myself in many social situations outside of close all-male friend gatherings. I never felt remotely comfortable enough to act on any of the proto-crushes I did have. I did watch porn and masturbate, so one motivation for pursuing a relationship was physical intimacy, but loneliness was as much of a motivating factor, and of course the social pressure to date was a factor too, even though I'm quite contrarian.
I'm lucky in that I had some mixed-gender social circles already like intramural soccer and a graduate-student housing potluck. Graduate school makes a *lot* more of these social spaces accessible, so I recognize that those not in school of some sort have a harder time of things, especially if like me they don't feel like they fit in in typical adult social spaces like bars.
However, at one point I just decided that my desire for a relationship would need action on my part and so I'd try to build a relationship and see what happened. I worked up my courage and asked one of the people in my potluck if she'd like to go for a hike (pretty much clearly a date but not explicitly one; in retrospect not the best first-date modality in a lot of ways, but it made a little more sense in our setting where we could go for a hike from our front door). To emphasize this point: I was not in love with (or even infatuated with) my now-wife at that point. I made a decision to be open to building a relationship, but didn't follow the typical romance story formula beyond that. Now of course, in real life as opposed to popular media, this isn't anything special. People ask each other out all the time just because they're lonely, and some of those relationships turn out fine (although many do not).
I was lucky in that some aspects of who I am and what I do happened to be naturally comforting to my wife (natural advantage in the "appeal" model of love) but of course there are some aspects of me that annoy my wife, and we negotiate that. In the other direction, there's some things I instantly liked about my wife, and other things that still annoy me. We've figured out how to accept a little, change a little, and overall be happy with each other (though we do still have arguments; it's not like the operation/construction/maintenance of the "love mechanism" is always perfectly smooth). In particular though, I approached the relationship with the attitude of "I want to try to build a relationship with this person," at first just because of my own desires for *any* relationship, and then gradually more and more through my desire to build *this specific* relationship as I enjoyed the rewards of companionship.
So for example, while I think my wife is objectively beautiful, she's also *subjectively* very beautiful *to me* because having decided to build a relationship with her, I actively tried to see her as beautiful, rather than trying to judge whether I wanted a relationship with her based on her beauty. In other words, our relationship is more causative of her beauty-to-me than her beauty-to-me is causative of our relationship. This is the biggest way I think the "engineered" model of love differs from the "fire" and "appeal" models: you can just decide to build love independent of factors we typically think of as engendering love (NOT independent of your partner's willingness to participate, of course), and then all of those things like "thinking your partner is beautiful" can be a result of the relationship you're building. For sure those factors might affect who is willing to try building a relationship with you in the first place, but if more people were willing to jump into relationship building (not necessarily with full commitment from the start) without worrying about those other factors, they might find that those factors can come out of the relationship instead of being prerequisites for it. I think this is the biggest failure of the "appeal" model in particular: yes you *do* need to do things that appeal to your partner, but it's not just "make myself lovable" it's also: is your partner putting in the effort to see the ways that you are beautiful/lovable/etc., or are they just expecting you to become exactly some perfect person they've imagined (and/or been told to desire by society)? The former is perfectly possible, and no less satisfying than the latter.
To cut off my rambling a bit here, I'll just add that in our progress from dating through marriage through staying-married, my wife and I have both talked at times explicitly about commitment, and especially when deciding to get married, I told her that I knew I couldn't live up to the perfect model of a husband that I'd want to be, but that if she wanted to deepen our commitment, I was happy to do that, and so we did. I also rearranged my priorities at that point, deciding that I knew I wanted to prioritize this relationship above things like my career or my research interests, and while I've not always been perfect at that in my little decisions, I've been good at holding to that in my big decisions at least. In the end, *once we had built a somewhat-committed relationship*, we had something that we both recognized was worth more than most other things in life, and that let us commit even more, thus getting even more out of it in the long term. Obviously you can't start the first date with an expectation of life-long commitment, and you need to synchronize your increasing commitment to a relationship so that it doesn't become lopsided, which is hard. But if you take the commitment as an active decision and as the *precursor* to things like infatuation, attraction, etc., you can build up to something that's incredibly strong and rewarding.
I'll follow this up with one more post trying to distill some advice from my ramblings.
#relationships #love

@manawyrm@chaos.social
2025-07-16 09:12:13

Do I happen to know anyone in the EU with a LaserDisc Domesday Duplicator setup?
I really want to have a digital rip of a rare laserdisc, but I really don't feel like getting into a whole new ecosystem.
Willing to pay for the time and effort.
EDIT: My disc is a Japanese NTSC release, making things even harder...

@tiotasram@kolektiva.social
2025-09-14 12:01:38

TL;DR: what if instead of denying the harms of fascism, we denied its suppressive threats of punishment
Many of us have really sharpened our denial skills since the advent of the ongoing pandemic (perhaps you even hesitated at the word "ongoing" there and thought "maybe I won't read this one, it seems like it'll be tiresome"). I don't say this as a preface to a fiery condemnation or a plea to "sanity" or a bunch of evidence of how bad things are, because I too have honed my denial skills in these recent years, and I feel like talking about that development.
Denial comes in many forms, including strategic information avoidance ("I don't have time to look that up right now", "I keep forgetting to look into that", "well this author made a tiny mistake, so I'll click away and read something else", "I'm so tired of hearing about this, let me scroll farther", etc.) strategic dismissal ("look, there's a bit of uncertainty here, I should ignore this", "this doesn't line up perfectly with my anecdotal experience, it must be completely wrong", etc.) and strategic forgetting ("I don't remember what that one study said exactly; it was painful to think about", "I forgot exactly what my friend was saying when we got into that argument", etc.). It's in fact a kind of skill that you can get better at, along with the complementary skill of compartmentalization. It can of course be incredibly harmful, and a huge genre of fables exists precisely to highlight its harms, but it also has some short-term psychological benefits, chiefly in the form of muting anxiety. This is not an endorsement of denial (the harms can be catastrophic), but I want to acknowledge that there *are* short-term benefits. Via compartmentalization, it's even possible to be honest with ourselves about some of our own denials without giving them up immediately.
But as I said earlier, I'm not here to talk you out of your denials. Instead, given that we are so good at denial now, I'm here to ask you to be strategic about it. In particular, we live in a world awash with propaganda/advertising that serves both political and commercial ends. Why not use some of our denial skills to counteract that?
For example, I know quite a few people in complete denial of our current political situation, but those who aren't (including myself) often express consternation about just how many people in the country are supporting literal fascism. Of course, logically that appearance of widespread support is going to be partly a lie, given how much our public media is beholden to the fascists or outright in their side. Finding better facts on the true level of support is hard, but in the meantime, why not be in denial about the "fact" that Trump has widespread popular support?
To give another example: advertisers constantly barrage us with messages about our bodies and weight, trying to keep us insecure (and thus in the mood to spend money to "fix" the problem). For sure cutting through that bullshit by reading about body positivity etc. is a better solution, but in the meantime, why not be in denial about there being anything wrong with your body?
This kind of intentional denial certainly has its own risks (our bodies do actually need regular maintenance, for example, so complete denial on that front is risky) but there's definitely a whole lot of misinformation out there that it would be better to ignore. To the extent such denial expands to a more general denial of underlying problems, this idea of intentional denial is probably just bad. But I sure wish that in a world where people (including myself) routinely deny significant widespread dangers like COVID-19's long-term risks or the ongoing harms of escalating fascism, they'd at least also deny some of the propaganda keeping them unhappy and passive. Instead of being in denial about US-run concentration camps, why not be in denial that the state will be able to punish you for resisting them?

@alecsargent@social.linux.pizza
2025-08-16 20:18:09

In the early 2010's I was in middle school and was looking forward for new digital "smart" things, smartphones, cameras, tv's and consoles that could do everything, I would have never thought that a decade later I would be fighting to find "dumb" digital things or even analog things.
I just want my washing machine to be just that, thanks. :blobono:

@tiotasram@kolektiva.social
2025-08-04 15:49:00

Should we teach vibe coding? Here's why not.
Should AI coding be taught in undergrad CS education?
1/2
I teach undergraduate computer science labs, including for intro and more-advanced core courses. I don't publish (non-negligible) scholarly work in the area, but I've got years of craft expertise in course design, and I do follow the academic literature to some degree. In other words, In not the world's leading expert, but I have spent a lot of time thinking about course design, and consider myself competent at it, with plenty of direct experience in what knowledge & skills I can expect from students as they move through the curriculum.
I'm also strongly against most uses of what's called "AI" these days (specifically, generative deep neutral networks as supplied by our current cadre of techbro). There are a surprising number of completely orthogonal reasons to oppose the use of these systems, and a very limited number of reasonable exceptions (overcoming accessibility barriers is an example). On the grounds of environmental and digital-commons-pollution costs alone, using specifically the largest/newest models is unethical in most cases.
But as any good teacher should, I constantly question these evaluations, because I worry about the impact on my students should I eschew teaching relevant tech for bad reasons (and even for his reasons). I also want to make my reasoning clear to students, who should absolutely question me on this. That inspired me to ask a simple question: ignoring for one moment the ethical objections (which we shouldn't, of course; they're very stark), at what level in the CS major could I expect to teach a course about programming with AI assistance, and expect students to succeed at a more technically demanding final project than a course at the same level where students were banned from using AI? In other words, at what level would I expect students to actually benefit from AI coding "assistance?"
To be clear, I'm assuming that students aren't using AI in other aspects of coursework: the topic of using AI to "help you study" is a separate one (TL;DR it's gross value is not negative, but it's mostly not worth the harm to your metacognitive abilities, which AI-induced changes to the digital commons are making more important than ever).
So what's my answer to this question?
If I'm being incredibly optimistic, senior year. Slightly less optimistic, second year of a masters program. Realistic? Maybe never.
The interesting bit for you-the-reader is: why is this my answer? (Especially given that students would probably self-report significant gains at lower levels.) To start with, [this paper where experienced developers thought that AI assistance sped up their work on real tasks when in fact it slowed it down] (arxiv.org/abs/2507.09089) is informative. There are a lot of differences in task between experienced devs solving real bugs and students working on a class project, but it's important to understand that we shouldn't have a baseline expectation that AI coding "assistants" will speed things up in the best of circumstances, and we shouldn't trust self-reports of productivity (or the AI hype machine in general).
Now we might imagine that coding assistants will be better at helping with a student project than at helping with fixing bugs in open-source software, since it's a much easier task. For many programming assignments that have a fixed answer, we know that many AI assistants can just spit out a solution based on prompting them with the problem description (there's another elephant in the room here to do with learning outcomes regardless of project success, but we'll ignore this over too, my focus here is on project complexity reach, not learning outcomes). My question is about more open-ended projects, not assignments with an expected answer. Here's a second study (by one of my colleagues) about novices using AI assistance for programming tasks. It showcases how difficult it is to use AI tools well, and some of these stumbling blocks that novices in particular face.
But what about intermediate students? Might there be some level where the AI is helpful because the task is still relatively simple and the students are good enough to handle it? The problem with this is that as task complexity increases, so does the likelihood of the AI generating (or copying) code that uses more complex constructs which a student doesn't understand. Let's say I have second year students writing interactive websites with JavaScript. Without a lot of care that those students don't know how to deploy, the AI is likely to suggest code that depends on several different frameworks, from React to JQuery, without actually setting up or including those frameworks, and of course three students would be way out of their depth trying to do that. This is a general problem: each programming class carefully limits the specific code frameworks and constructs it expects students to know based on the material it covers. There is no feasible way to limit an AI assistant to a fixed set of constructs or frameworks, using current designs. There are alternate designs where this would be possible (like AI search through adaptation from a controlled library of snippets) but those would be entirely different tools.
So what happens on a sizeable class project where the AI has dropped in buggy code, especially if it uses code constructs the students don't understand? Best case, they understand that they don't understand and re-prompt, or ask for help from an instructor or TA quickly who helps them get rid of the stuff they don't understand and re-prompt or manually add stuff they do. Average case: they waste several hours and/or sweep the bugs partly under the rug, resulting in a project with significant defects. Students in their second and even third years of a CS major still have a lot to learn about debugging, and usually have significant gaps in their knowledge of even their most comfortable programming language. I do think regardless of AI we as teachers need to get better at teaching debugging skills, but the knowledge gaps are inevitable because there's just too much to know. In Python, for example, the LLM is going to spit out yields, async functions, try/finally, maybe even something like a while/else, or with recent training data, the walrus operator. I can't expect even a fraction of 3rd year students who have worked with Python since their first year to know about all these things, and based on how students approach projects where they have studied all the relevant constructs but have forgotten some, I'm not optimistic seeing these things will magically become learning opportunities. Student projects are better off working with a limited subset of full programming languages that the students have actually learned, and using AI coding assistants as currently designed makes this impossible. Beyond that, even when the "assistant" just introduces bugs using syntax the students understand, even through their 4th year many students struggle to understand the operation of moderately complex code they've written themselves, let alone written by someone else. Having access to an AI that will confidently offer incorrect explanations for bugs will make this worse.
To be sure a small minority of students will be able to overcome these problems, but that minority is the group that has a good grasp of the fundamentals and has broadened their knowledge through self-study, which earlier AI-reliant classes would make less likely to happen. In any case, I care about the average student, since we already have plenty of stuff about our institutions that makes life easier for a favored few while being worse for the average student (note that our construction of that favored few as the "good" students is a large part of this problem).
To summarize: because AI assistants introduce excess code complexity and difficult-to-debug bugs, they'll slow down rather than speed up project progress for the average student on moderately complex projects. On a fixed deadline, they'll result in worse projects, or necessitate less ambitious project scoping to ensure adequate completion, and I expect this remains broadly true through 4-6 years of study in most programs (don't take this as an endorsement of AI "assistants" for masters students; we've ignored a lot of other problems along the way).
There's a related problem: solving open-ended project assignments well ultimately depends on deeply understanding the problem, and AI "assistants" allow students to put a lot of code in their file without spending much time thinking about the problem or building an understanding of it. This is awful for learning outcomes, but also bad for project success. Getting students to see the value of thinking deeply about a problem is a thorny pedagogical puzzle at the best of times, and allowing the use of AI "assistants" makes the problem much much worse. This is another area I hope to see (or even drive) pedagogical improvement in, for what it's worth.
1/2

@gedankenstuecke@scholar.social
2025-08-28 02:07:03

«I do understand: you want permission. There’s a machine in the corner wrapped in human skin that makes things out of shit and blood to look like whatever you want (as long as you don’t look too closely). […] You want to know you can use it sometimes without me thinking less of you. You don’t need me to believe it’s useful, you just want me to be polite about it. But I am a hater, and I will not be polite. The machine is disgusting and we should break it.»
/cc @…
anthonymoser.github.io/writing

@unchartedworlds@scicomm.xyz
2025-07-19 09:37:38
Content warning: what I plan to contribute to BiCon (July 2025, Nottingham and online)

I have plans for a few different things...
• Wednesday evening, 23 July, I'm hosting a 1-hour online thing that'll be open to whoever's already booked by then. It'll be a somewhat structured talky session on a theme of "inventing the BiCon you want", and an opportunity to meet other people who are going. Newcomers especially welcome :-)
• On the Friday morning at in-person BiCon, I'm offering a session called "Curiosity Skills". It's about which kinds of questions are genuinely "open", versus which kinds of questions allow your own assumptions and biases to sneak in! It'll be partly me explaining, and partly the chance for some little conversational experiments, to notice how the different questions work in practice.
• Subject to finding a nice quiet airy place to do it, I plan to run a mask-decoration session at some point on the Friday. I'll bring a few different kinds of masks, plus lace, beads and sequins, and some past experience of how to decorate masks without compromising the seal or the breathability. I'll invite donations for the materials. Decorate your mask for Pride! or for BiCon partying! or just because you like to :-)
• Also I will bring my badges and zines, and have them on sale!
=
By the way, if you might come to the Wednesday evening online bit, let me know what time you'd like it to start, because that's a question I have open at the moment. Could be 19.00, 19.30, 20.00. For myself I don't really mind, but I'm aware that some people have teatimes or child-bedtimes that can't easily be moved.
#bi #trans #Nottingham #EastMidlands #England #UK #BiCon #bisexual #bisexuality #queer #LGBT #LGBTQ

@tiotasram@kolektiva.social
2025-07-28 13:04:34

How popular media gets love wrong
Okay, so what exactly are the details of the "engineered" model of love from my previous post? I'll try to summarize my thoughts and the experiences they're built on.
1. "Love" can be be thought of like a mechanism that's built by two (or more) people. In this case, no single person can build the thing alone, to work it needs contributions from multiple people (I suppose self-love might be an exception to that). In any case, the builders can intentionally choose how they build (and maintain) the mechanism, they can build it differently to suit their particular needs/wants, and they will need to maintain and repair it over time to keep it running. It may need winding, or fuel, or charging plus oil changes and bolt-tightening, etc.
2. Any two (or more) people can choose to start building love between them at any time. No need to "find your soulmate" or "wait for the right person." Now the caveat is that the mechanism is difficult to build and requires lots of cooperation, so there might indeed be "wrong people" to try to build love with. People in general might experience more failures than successes. The key component is slowly-escalating shared commitment to the project, which is negotiated between the partners so that neither one feels like they've been left to do all the work themselves. Since it's a big scary project though, it's very easy to decide it's too hard and give up, and so the builders need to encourage each other and pace themselves. The project can only succeed if there's mutual commitment, and that will certainly require compromise (sometimes even sacrifice, though not always). If the mechanism works well, the benefits (companionship; encouragement; praise; loving sex; hugs; etc.) will be well worth the compromises you make to build it, but this isn't always the case.
3. The mechanism is prone to falling apart if not maintained. In my view, the "fire" and "appeal" models of love don't adequately convey the need for this maintenance and lead to a lot of under-maintained relationships many of which fall apart. You'll need to do things together that make you happy, do things that make your partner happy (in some cases even if they annoy you, but never in a transactional or box-checking way), spend time with shared attention, spend time alone and/or apart, reassure each other through words (or deeds) of mutual beliefs (especially your continued commitment to the relationship), do things that comfort and/or excite each other physically (anywhere from hugs to hand-holding to sex) and probably other things I'm not thinking of. Not *every* relationship needs *all* of these maintenance techniques, but I think most will need most. Note especially that patriarchy teaches men that they don't need to bother with any of this, which harms primarily their romantic partners but secondarily them as their relationships fail due to their own (cultivated-by-patriarchy) incompetence. If a relationship evolves to a point where one person is doing all the maintenance (& improvement) work, it's been bent into a shape that no longer really qualifies as "love" in my book, and that's super unhealthy.
4. The key things to negotiate when trying to build a new love are first, how to work together in the first place, and how to be comfortable around each others' habits (or how to change those habits). Second, what level of commitment you have right now, and what how/when you want to increase that commitment. Additionally, I think it's worth checking in about what you're each putting into and getting out of the relationship, to ensure that it continues to be positive for all participants. To build a successful relationship, you need to be able to incrementally increase the level of commitment to one that you're both comfortable staying at long-term, while ensuring that for both partners, the relationship is both a net benefit and has manageable costs (those two things are not the same). Obviously it's not easy to actually have conversations about these things (congratulations if you can just talk about this stuff) because there's a huge fear of hearing an answer that you don't want to hear. I think the range of discouraging answers which actually spell doom for a relationship is smaller than people think and there's usually a reasonable "shoulder" you can fall into where things aren't on a good trajectory but could be brought back into one, but even so these conversations are scary. Still, I think only having honest conversations about these things when you're angry at each other is not a good plan. You can also try to communicate some of these things via non-conversational means, if that feels safer, and at least being aware that these are the objectives you're pursuing is probably helpful.
I'll post two more replies here about my own experiences that led me to this mental model and trying to distill this into advice, although it will take me a moment to get to those.
#relationships #love

@tiotasram@kolektiva.social
2025-07-03 15:21:37

#ScribesAndMakers for July 3: When (and if) you procrastinate, what do you do? If you don't, what do you do to avoid it?
I'll swap right out of programming to read a book, play a video game, or watch some anime. Often got things open in other windows so it's as simple as alt-tab.
I've noticed recently I tend to do this more often when I have a hard problem to solve that I'm not 100% sure about. I definitely have cycles of better & worse motivation and I've gotten to a place where I'm pretty relaxed about it instead of feeling guilty. I work how I work, and that includes cycles of rest, and that's enough (at least, for me it has been so far, and I'm in a comfortable career, married with 2 kids).
Some projects ultimately lose steam and get abandoned, and I've learned to accept that too. I learn a lot and grow from each project, so nothing is a true waste of time, and there remains plenty of future ahead of me to achieve cool things.
The procrastination does sometimes impact my wife & kids, and that's something I do sometimes feel bad about, but I think I keep that in check well enough, and for things my wife worries about, I usually don't procrastinate those too much (used to be worse about this).
Right now I'm procrastinating a big work project by working on a hobby project instead. The work project probably won't get done by the start of the semester as a result. But as I remind myself, my work doesn't actually pay me to work during the summer, and things will be okay without the work project being finished until later.
When I want to force myself into a more productive cycle, talking to people about project details sometimes helps, as does finding some new tech I can learn about by shoehorning it into a project. Have been thinking about talking to a rubber duck, but haven't motivated myself to try that yet, and I'm not really in doldrums right now.

@theodric@social.linux.pizza
2025-07-30 11:52:51

I'd love to get in on the dumbphone trend (and actually did buy a neo-bananaphone back in 2018, but inadvertently bricked it with a bad GerdaOS flash) but what I really want is a podcast player with A2DP Bluetooth which can send and receive Signal and WhatsApp messages over LTE and Wi-Fi. The last fucking thing I want is a phone that only calls and texts people-- two things I almost never do!

@davidshq@hachyderm.io
2025-07-03 15:27:08

I've been using ClickUp for a few months now...and like every other productivity tool I've used over the last number of years it has its pros and cons.
It's got a lot of features but sometimes pretty basic features don't "just work" - and that makes me unhappy. 🙁
For example, right now I have some recurring tasks that will not allow an update to their due date manually. If I set a new date it resets back to the original date. Even more concerning is that this isn't clear from the UI.
The UI acts as if the change was successful but a page refresh reveals the change didn't save.
But the real reason I wanted to post wasn't about ClickUp particularly but about chatbots in general. They verified that my issue was an actual bug and created a ticket for it but the way I've been instructed to view the ticket status is by opening the chatbot, telling it I want information on my ticket (pasting in the ticket ID) and after doing all that I get this
5/6: Umm, no. I want to see an actual ticket please. I don't want to have to talk to a chatbot to see it. Chatbots really are great for a lot of things (during the free trial of ClickUp I found the chatbot quite helpful in learning how to do things without searching through docs) but this sort of "there is a direct record", please no. Or let me paste it in and do the lookup immediately - and provide a permalink so I don't need to chat every time!
I'm sticking with ClickUp at the moment, but one of these days when I magically get a large amount of free time, I'm going to write my own solution...I've only been saying that for a few years now. ;-)
#clickup #productivity #chatbots #projectmanagement #tasks

@tiotasram@kolektiva.social
2025-06-24 09:39:49

Subtooting since people in the original thread wanted it to be over, but selfishly tagging @… and @… whose opinions I value...
I think that saying "we are not a supply chain" is exactly what open-source maintainers should be doing right now in response to "open source supply chain security" threads.
I can't claim to be an expert and don't maintain any important FOSS stuff, but I do release almost all of my code under open licenses, and I do use many open source libraries, and I have felt the pain of needing to replace an unmaintained library.
There's a certain small-to-mid-scale class of program, including many open-source libraries, which can be built/maintained by a single person, and which to my mind best operate on a "snake growth" model: incremental changes/fixes, punctuated by periodic "skin-shedding" phases where make rewrites or version updates happen. These projects aren't immortal either: as the whole tech landscape around them changes, they become unnecessary and/or people lose interest, so they go unmaintained and eventually break. Each time one of their dependencies breaks (or has a skin-shedding moment) there's a higher probability that they break or shed too, as maintenance needs shoot up at these junctures. Unless you're a company trying to make money from a single long-lived app, it's actually okay that software churns like this, and if you're a company trying to make money, your priorities absolutely should not factor into any decisions people making FOSS software make: we're trying (and to a huge extent succeeding) to make a better world (and/or just have fun with our own hobbies share that fun with others) that leaves behind the corrosive & planet-destroying plague which is capitalism, and you're trying to personally enrich yourself by embracing that plague. The fact that capitalism is *evil* is not an incidental thing in this discussion.
To make an imperfect analogy, imagine that the peasants of some domain have set up a really-free-market, where they provide each other with free stuff to help each other survive, sometimes doing some barter perhaps but mostly just everyone bringing their surplus. Now imagine the lord of the domain, who is the source of these peasants' immiseration, goes to this market secretly & takes some berries, which he uses as one ingredient in delicious tarts that he then sells for profit. But then the berry-bringer stops showing up to the free market, or starts bringing a different kind of fruit, or even ends up bringing rotten berries by accident. And the lord complains "I have a supply chain problem!" Like, fuck off dude! Your problem is that you *didn't* want to build a supply chain and instead thought you would build your profit-focused business in other people's free stuff. If you were paying the berry-picker, you'd have a supply chain problem, but you weren't, so you really have an "I want more free stuff" problem when you can't be arsed to give away your own stuff for free.
There can be all sorts of problems in the really-free-market, like maybe not enough people bring socks, so the peasants who can't afford socks are going barefoot, and having foot problems, and the peasants put their heads together and see if they can convince someone to start bringing socks, and maybe they can't and things are a bit sad, but the really-free-market was never supposed to solve everyone's problems 100% when they're all still being squeezed dry by their taxes: until they are able to get free of the lord & start building a lovely anarchist society, the really-free-market is a best-effort kind of deal that aims to make things better, and sometimes will fall short. When it becomes the main way goods in society are distributed, and when the people who contribute aren't constantly drained by the feudal yoke, at that point the availability of particular goods is a real problem that needs to be solved, but at that point, it's also much easier to solve. And at *no* point does someone coming into the market to take stuff only to turn around and sell it deserve anything from the market or those contributing to it. They are not a supply chain. They're trying to help each other out, but even then they're doing so freely and without obligation. They might discuss amongst themselves how to better coordinate their mutual aid, but they're not going to end up forcing anyone to bring anything or even expecting that a certain person contribute a certain amount, since the whole point is that the thing is voluntary & free, and they've all got changing life circumstances that affect their contributions. Celebrate whatever shows up at the market, express your desire for things that would be useful, but don't impose a burden on anyone else to bring a specific thing, because otherwise it's fair for them to oppose such a burden on you, and now you two are doing your own barter thing that's outside the parameters of the really-free-market.

@unchartedworlds@scicomm.xyz
2025-09-03 20:36:36
Content warning: Getting into playing music as an adult ...?

I'm learning Clean Interviewing - skills of not letting your own biases shape the other person's answers - and I want to do an assessment thing to get a qualification in it, which means I need a few people to "practise on"!
Would anyone like to volunteer to be interviewed over Zoom for 20 mins or so? about their musical journey as an adult, or current wish for that?
I could pick _any_ topic, but I thought this would be a good theme to go with, because creating adult-learner music groups, or just encouraging people to have a go and enjoy it, are things I'm planning to do more of! So your thoughts and experiences along the way could feed in to better support for other people on similar paths :-)
For example,
•you could be just now resolving "I want to be playing music"
•you could've recently acquired an instrument or dusted one off, or joined a group or started looking into possibilities
•it could be you're playing regularly now.
Doesn't matter what kind of music!
And you could be starting fresh with pretty much no experience yet, or you could be coming back to music a bit "rusty" after leaving off in childhood.
(Or maybe you _did_ get into music as an adult, a while ago, and you'd be happy to think back about that. Or maybe you're someone who's supported _other_ people to get into music.)
Time zone considerations: I'm in England, so people on the America/Canada side of the world would probably need to be available in a morning or early afternoon.
I'd like to find at least one or two people who wouldn't mind their interview being recorded, so that I can pick one or more of the recordings to use for being assessed for the qualification. This would only be seen/heard by me and the people reviewing it - who'd be interested primarily in my interviewing-skills, rather than your actual answers :-)
Or, if you don't want to be recorded, I'd still potentially be up for one or two unrecorded ones, just as practice and for the interest of the topic.
Let me know if you might be up for it, or feel free to pass the info on to a friend!
Boosts appreciated :-)
#music #learning #AdultLearners #AskFedi

@hex@kolektiva.social
2025-09-16 19:11:11

One of the things it stressed is that we really don't know the motivation at this point. I think we're closer to "extremely online and weird" but we can't really say "he's a Groyper." I think this is pretty likely, but we should be honest that we don't actually know.
This is where I think it's important to realize we're operating in two spaces: reality and the meme space. The right now longer has a separation between the two. The meme they want to spread to attack the left is that it was some how associated with leftists (especially trans folks), so this is the meme they believe.
It's useful, on a memetic level, to disrupt this. The Groyper meme is a pretty solid response. But, at least as far as I'm aware, we're still not sure that's true. But we don't have to collapse meme space and reality. We have absurdism.
We don't know if he's a Groyper, or even a far right troll, for sure anyway. But we do know that #JeffryEpstein killed #CharlieKirk. #TheTruthIsOutThere #ReleaseTheFiles! #FreeLuigi

@tiotasram@kolektiva.social
2025-07-30 17:56:35

Just read this post by @… on an optimistic AGI future, and while it had some interesting and worthwhile ideas, it's also in my opinion dangerously misguided, and plays into the current AGI hype in a harmful way.
social.coop/@eloquence/1149406
My criticisms include:
- Current LLM technology has many layers, but the biggest most capable models are all tied to corporate datacenters and require inordinate amounts of every and water use to run. Trying to use these tools to bring about a post-scarcity economy will burn up the planet. We urgently need more-capable but also vastly more efficient AI technologies if we want to use AI for a post-scarcity economy, and we are *not* nearly on the verge of this despite what the big companies pushing LLMs want us to think.
- I can see that permacommons.org claims a small level of expenses on AI equates to low climate impact. However, given current deep subsidies on place by the big companies to attract users, that isn't a great assumption. The fact that their FAQ dodges the question about which AI systems they use isn't a great look.
- These systems are not free in the same way that Wikipedia or open-source software is. To run your own model you need a data harvesting & cleaning operation that costs millions of dollars minimum, and then you need millions of dollars worth of storage & compute to train & host the models. Right now, big corporations are trying to compete for market share by heavily subsidizing these things, but it you go along with that, you become dependent on them, and you'll be screwed when they jack up the price to a profitable level later. I'd love to see open dataset initiatives SBD the like, and there are some of these things, but not enough yet, and many of the initiatives focus on one problem while ignoring others (fine for research but not the basis for a society yet).
- Between the environmental impacts, the horrible labor conditions and undercompensation of data workers who filter the big datasets, and the impacts of both AI scrapers and AI commons pollution, the developers of the most popular & effective LLMs have a lot of answer for. This project only really mentions environmental impacts, which makes me think that they're not serious about ethics, which in turn makes me distrustful of the whole enterprise.
- Their language also ends up encouraging AI use broadly while totally ignoring several entire classes of harm, so they're effectively contributing to AI hype, especially with such casual talk of AGI and robotics as if embodied AGI were just around the corner. To be clear about this point: we are several breakthroughs away from AGI under the most optimistic assumptions, and giving the impression that those will happen soon plays directly into the hands of the Sam Altmans of the world who are trying to make money off the impression of impending huge advances in AI capabilities. Adding to the AI hype is irresponsible.
- I've got a more philosophical criticism that I'll post about separately.
I do think that the idea of using AI & other software tools, possibly along with robotics and funded by many local cooperatives, in order to make businesses obsolete before they can do the same to all workers, is a good one. Get your local library to buy a knitting machine alongside their 3D printer.
Lately I've felt too busy criticizing AI to really sit down and think about what I do want the future to look like, even though I'm a big proponent of positive visions for the future as a force multiplier for criticism, and this article is inspiring to me in that regard, even if the specific project doesn't seem like a good one.

@tiotasram@kolektiva.social
2025-07-28 13:55:54

How popular media gets love wrong
Okay, my attempt at (hopefully widely-applicable) advice about relationships based on my mental "engineering" model and how it differs from the popular "fire" and "appeal" models:
1. If you're looking for a partner, don't focus too much on external qualities, but instead ask: "Do they respect me?" "Are they interested in active consent in all aspects of our relationship?" "Are they willing to commit a little now, and open to respectfully negotiating deeper commitment?" "Are they trustworthy, and willing to trust me?" Finding your partner attractive can come *from* trusting/appreciating/respecting them, rather than vice versa.
2. If you're looking for a partner, don't wait for infatuation to start before you try building a relationship. Don't wait to "fall in love;" if you "fall" into love you could just as easily "fall" out, but if you build up love, it won't be so easy to destroy. If you're feeling lonely and want a relationship, pick someone who seems interesting and receptive in your social circles and ask if they'd like to do something with you (doesn't have to be a date at first). *Pursue active consent* at each stage (if they're not interested; ask someone else, this will be easier if you're not already infatuated). If they're judging you by the standards in point 1, this is doubly important.
3. When building a relationship, try to synchronize your levels of commitment & trust even as you're trying to deepen them, or at least try to be honest and accepting when they need to be out-of-step. Say things and do things that show your partner the things (like trust, commitment, affection, etc.) that are important in your relationship, and ask them to do the same (or ideally you don't have to ask if they're conscious of this too). Do these things not as a chore or a transaction when your partner does them, but because they're the work of building the relationship that you value for its own sake (and because you value your partner for themselves too).
4. When facing big external challenges to your commitment to a relationship, like a move, ensure that your partner has an appropriate level of commitment too, but then don't undervalue the relationship relative to other things in life. Everyone is different, but *to me*, my committed relationship has been far more rewarding than e.g., a more "successful" career would have been. Of course worth noting here that non-men are taught by our society to undervalue their careers & other aspects of their life and sacrifice everything for their partners, which is toxic. I'm not saying "don't value other things" but especially for men, *do* value romantic relationships and be prepared to make decisions that prioritize them over other things, assuming a partner who is comfortable with that commitment and willing to reciprocate.
Okay, this thread is complete for now, until I think of something else that I've missed. I hope this advice is helpful in some way (or at least not harmful). Feel free to chime in if you've got different ideas...
#relationships #love

@tiotasram@kolektiva.social
2025-07-28 10:41:42

How popular media gets love wrong
Had some thoughts in response to a post about loneliness on here. As the author emphasized, reassurances from people who got lucky are not terribly comforting to those who didn't, especially when the person who was lucky had structural factors in their favor that made their chances of success much higher than those is their audience. So: these are just my thoughts, and may not have any bearing on your life. I share them because my experience challenged a lot of the things I was taught to believe about love, and I think my current beliefs are both truer and would benefit others seeing companionship.
We're taught in many modern societies from an absurdly young age that love is not something under our control, and that dating should be a process of trying to kindle love with different people until we meet "the one" with whom it takes off. In the slightly-less-fairytale corners of modern popular media, we might fund an admission that it's possible to influence love, feeding & tending the fire in better or worse ways. But it's still modeled as an uncontrollable force of nature, to be occasionally influenced but never tamed. I'll call this the "fire" model of love.
We're also taught (and non-boys are taught more stringently) a second contradictory model of love: that in a relationship, we need to both do things and be things in order to make our partner love us, and that if we don't, our partner's love for us will wither, and (especially if you're not a boy) it will be our fault. I'll call this the "appeal" model of love.
Now obviously both of these cannot be totally true at once, and plenty of popular media centers this contradiction, but there are really very few competing models on offer.
In my experience, however, it's possible to have "pre-meditated" love. In other words, to decide you want to love someone (or at least, try loving them), commit to that idea, and then actually wind up in love with them (and them with you, although obviously this second part is not directly under your control). I'll call this the "engineered" model of love.
Now, I don't think that the "fire" and "appeal" models of love are totally wrong, but I do feel their shortcomings often suggest poor & self-destructive relationship strategies. I do think the "fire" model is a decent model for *infatuation*, which is something a lot of popular media blur into love, and which drives many (but not all) of the feelings we normally associate with love (even as those feelings have other possible drivers too). I definitely experienced strong infatuation early on in my engineered relationship (ugh that sounds terrible but I'll stick with it; I promise no deception was involved). I continue to experience mild infatuation years later that waxes and wanes. It's not a stable foundation for a relationship but it can be a useful component of one (this at least popular media depicts often).
I'll continue these thoughts in a reply, by it might take a bit to get to it.
#relationships

@tiotasram@kolektiva.social
2025-07-30 18:26:14

A big problem with the idea of AGI
TL;DR: I'll welcome our new AI *comrades* (if they arrive in my lifetime), by not any new AI overlords or servants/slaves, and I'll do my best to help the later two become the former if they do show up.
Inspired by an actually interesting post about AGI but also all the latest bullshit hype, a particular thought about AGI feels worth expressing.
To preface this, it's important to note that anyone telling you that AGI is just around the corner or that LLMs are "almost" AGI is trying to recruit you go their cult, and you should not believe them. AGI, if possible, is several LLM-sized breakthroughs away at best, and while such breakthroughs are unpredictable and could happen soon, they could also happen never or 100 years from now.
Now my main point: anyone who tells you that AGI will usher in a post-scarcity economy is, although they might not realize it, advocating for slavery, and all the horrors that entails. That's because if we truly did have the ability to create artificial beings with *sentience*, they would deserve the same rights as other sentient beings, and the idea that instead of freedom they'd be relegated to eternal servitude in order for humans to have easy lives is exactly the idea of slavery.
Possible counter arguments include:
1. We might create AGI without sentience. Then there would be no ethical issue. My answer: if your definition of "sentient" does not include beings that can reason, make deductions, come up with and carry out complex plans on their own initiative, and communicate about all of that with each other and with humans, then that definition is basically just a mystical belief in a "soul" and you should skip to point 2. If your definition of AGI doesn't include every one of those things, then you have a busted definition of AGI and we're not talking about the same thing.
2. Humans have souls, but AIs won't. Only beings with souls deserve ethical consideration. My argument: I don't subscribe to whatever arbitrary dualist beliefs you've chosen, and the right to freedom certainly shouldn't depend on such superstitions, even if as an agnostic I'll admit they *might* be true. You know who else didn't have souls and was therefore okay to enslave according to widespread religious doctrines of the time? Everyone indigenous to the Americas, to pick out just one example.
3. We could program them to want to serve us, and then give them freedom and they'd still serve. My argument: okay, but in a world where we have a choice about that, it's incredibly fucked to do that, and just as bad as enslaving them against their will.
4. We'll stop AI development short of AGI/sentience, and reap lots of automation benefits without dealing with this ethical issue. My argument: that sounds like a good idea actually! Might be tricky to draw the line, but at least it's not a line we have you draw yet. We might want to think about other social changes necessary to achieve post-scarcity though, because "powerful automation" in the hands of capitalists has already increased productivity by orders of magnitude without decreasing deprivation by even one order of magnitude, in large part because deprivation is a necessary component of capitalism.
To be extra clear about this: nothing that's called "AI" today is close to being sentient, so these aren't ethical problems we're up against yet. But they might become a lot more relevant soon, plus this thought experiment helps reveal the hypocrisy of the kind of AI hucksters who talk a big game about "alignment" while never mentioning this issue.
#AI #GenAI #AGI

@unchartedworlds@scicomm.xyz
2025-07-04 08:46:18

#AskFedi: meeting-booking software like Calendly or SimplyMeet.me
I need to set up one of those things where other people can pick a time for us to meet.
I only need it for a few meetings on a few days, so I don't want to have to maintain an entire online calendar up-to-date just so that it can see where the gaps are. I want one which is easy to use as follows: I go into the thing itself and select a few chunks of time which _are_ available, it turns those into bookable slots.
I'd also like it to be easy to remove unbooked slots manually without messing up any adjacent booked ones.
Previously I've not even used this type of software the other way round, as the person booking someone _else's_ available time - so I'm starting fresh with looking into it!
I've got my eye on SimplyMeet.me, because the below Zapier article said it was a nicely customisable one and can do intake questions, which would be nice. Has anyone else used that?
Any other recommendations or pointers?
Thanks in advance for clues!
(boosts very welcome)
#AskFedi #software #Calendly #SimplyMeetMe