Searching We.Love.Privacy.Club

Twts matching #Ideas
Sort by: Newest, Oldest, Most Relevant

Gajim: Gajim 1.4 Preview: Workspaces
The Gajim team has been hard at work in the past months to prepare the next v1.4 release. The upcoming version brings a major interface redesign. In this post, we explain how the new interface works and what remains to be decided or implemented before the release.

Of course, your feedback is important! No interface can please everyone, so please react to this post with how this change would impact you positively and negatively, and ideas you have to make it even better … ⌘ Read more

⤋ Read More
In-reply-to » I wrote a 'banner'-like program for Plan 9 (and p9p) that uses the Unicode box drawing characters: http://txtpunk.com/banner/index.html

No, I’m still doing them manually. 🤣🤦🏻 But I do think they are a good idea and will be adding them, I just haven’t gotten around to finding a compatible implementation of the hash yet.

⤋ Read More
In-reply-to » My finger server now includes the last post from tw that doesn't have a subject. 'finger a@9srv.net'

With the finger server specifically? No idea, it’s a toy. I’d honestly forgotten I had it on until someone mentioned finger.farm and I was inspired to poke at it again.

⤋ Read More

@prologic@twtxt.net sounds about right. I tend to try to build my own before pulling in libs. learn more that way. I was looking at using it as a way to build my twt mirroring idea. and testing the lex parser with a wide ranging corpus to find edge cases. (the pgp signed feeds for one)

⤋ Read More

Would online dating without images lead to deeper, more human connections? I.e. only descriptions of people. If yes, is it different because of molochian reasons? More beautiful people have no problem showing their faces, so not showing ones face is seen as a low-status signal at some point. Counter: The idea of deeper, more human connections is in itself flawed, most mating choices are the result of a combination of class/status signals and physical attractiveness anyway.

⤋ Read More

@prologic@twtxt.net My thoughts on it being if they switched from a different way of hosting the file or multiple locations for redundancy..

I have an idea of using something like SRV records where they can define weighted url endpoints to reach.

⤋ Read More

I have trouble with a web crawler using the TOR network. It’s misusing the gopher proxy on my page. I don’t want to disable/block tor (that would be the easy way out). It’s permanently changing user agents and ignoring robots.txt. It ignores HTTP status codes. I’m currently serving it 4MB binary garbage in form of Link. It sucked in about 40GB of data now, but it doesn’t explode and keeps crawling. Any other idea about what to do with it?

⤋ Read More