Searching We.Love.Privacy.Club

Twts matching #processing
Sort by: Newest, Oldest, Most Relevant

ProcessOne: ejabberd 21.12
This new ejabberd 21.12 release comes after five months of work, contains more than one hundred changes, many of them are major improvements or features, and several bug fixes.

Image

When upgrading from previous versions, please notice: there’s a change in mod_register_web behaviour, and PosgreSQL database, please take a look if they affect your installation.

A more detailed expla … ⌘ Read more

⤋ Read More

ProcessOne: ejabberd 21.12
This new ejabberd 21.12 release comes after five months of work, contains more than one hundred changes, many of them are major improvements or features, and several bug fixes.

Image

When upgrading from previous versions, please notice: there’s a change in mod_register_web behaviour, and PosgreSQL database, please take a look if they affect your installation.

A more detailed expla … ⌘ Read more

⤋ Read More

Building Docker images in Drone CI using Docker-in-Docker
This evening I tried to improve the build process of GoBlog. GoBlog gets built using Drone CI and Docker. The problem was that two image variants are to be built, one based on the other, and the whole thing always took quite a long time. ⌘ Read more

⤋ Read More

Erlang Solutions: Blockchain Tech Deep Dive ¼

INTRODUCTION

Blockchain technology is transforming nearly every industry, whether it be banking,  government, fashion or logistics. The benefits of using blockchain are substantial – businesses can lower transaction costs, free up capital, speed up processes, and enhance security and trust. So it’s no surprise that more and more companies and developers are interested in working with the technology and leveraging its potential than ev … ⌘ Read more

⤋ Read More

Blue-teaming for Exiv2, part 1: creating a security advisory process
This blog post is the first in a series about hardening the security of the Exiv2 project. My goal is to share tips that will help you harden the security of your own project. ⌘ Read more

⤋ Read More

@movq@www.uninformativ.de What do you think about this?

diff –git a/jenny b/jenny
index b47c78e..20cf659 100755
— a/jenny
+++ b/jenny
@@ -278,7 +278,8 @@ def prefill_for(email, reply_to_this, self_mentions):
def process_feed(config, nick, url, content, lasttwt):

 nick_address, nick_desc = decide_nick(content, nick)
 url_for_hash = decide_url_for_hash(content, url)
  • new_lasttwt = parse(‘1800-01-01T12:00:00+00:00’).timestamp()

  • # new_lasttwt = parse(‘1800-01-01T12:00:00+00:00’).timestamp()

  • new_lasttwt = None

    for line in twt_lines_from_content(content):

     res = twt_line_to_mail(
    

    @@ -296,7 +297,7 @@ def process_feed(config, nick, url, content, lasttwt):

     twt_stamp = twt_date.timestamp()
     if lasttwt is not None and lasttwt >= twt_stamp:
         continue
    
  • if twt_stamp > new_lasttwt:

  • if not new_lasttwt or twt_stamp > new_lasttwt:

         new_lasttwt = twt_stamp
    
    
     mailname_new = join(config['maildir_target'], 'new', twt_hash)
    

⤋ Read More
In-reply-to » My thoughts about range requests

@movq@www.uninformativ.de

Don’t miss step 0 (I should have made this a separate point): having a meta header promising appending twts with strictly monotonically increasing timestamps.

(Also, I’d first like to see the pagination thingy implemented.)

In jenny I would like to see “don’t process previously fetched twts” AKA “Allow the user to archive/delete old twts” feature implemented ;-)

⤋ Read More

My thoughts about range requests

Additionally to pagination also range request should be used to reduce traffic.

I understand that there are corner cases making this a complicated matter.

I would like to see a meta header saying that the given twtxt is append only with increasing timestamps so that a simple strategy can detect valid content fetched per range request.

  1. read meta part per range request
  2. read last fetched twt at expected range (as known from last fetch)
  3. if fetched content starts with expected twt then process rest of data
  4. if fetched content doesn’t start with expected twt discard all and fall back to fetching whole twtxt

Pagination (e.g. archiving old content in a different file) will lead to point 4.

Of course especially pods should support range requests, correct @prologic@twtxt.net?

⤋ Read More

@movq@www.uninformativ.de I am getting this when I run it on cron (extra lines in between becuase otherwise jenny will make them a mash):

Traceback (most recent call last):

File “/home/quark/jenny/jenny”, line 565, in

if not retrieve_all(config):

File “/home/quark/jenny/jenny”, line 373, in retrieve_all

refresh_self(config)

File “/home/quark/jenny/jenny”, line 294, in refresh_self

process_feed(config, config[‘self_nick’], config[‘self_url’], content)

File “/home/quark/jenny/jenny”, line 280, in process_feed

fp.write(mail_body)

File “/usr/lib/python3.8/encodings/iso8859_15.py”, line 19, in encode

return codecs.charmap_encode(input,self.errors,encoding_table)[0]

UnicodeEncodeError: ‘charmap’ codec can’t encode character ‘\U0001f4e3’ in position 31: character maps to

⤋ Read More

An analysis on developer-security researcher interactions in the vulnerability disclosure process
We put out a call to open source developers and security researchers to talk about the security vulnerability disclosure process. Here’s what we found. ⌘ Read more

⤋ Read More

When tragedy strikes unexpectedly we cannot just go on as if nothing happened. Our minds need to be given time to deal with the blow. So it is necessary to pause and allow ourselves to process and recover.

⤋ Read More

@movq@www.uninformativ.de “Random thought: Would be great if you could do for i in ...; do something "$i" & done ; wait in a Shell script, but with the Shell only spawning one process per CPU.” -> Interesting which annoyances stay in the back of the head – I’d never articulated this, but it’s absolutely true that this would be great.

⤋ Read More

Urban Fish Ponds: Low-tech Sewage Treatment for Towns and Cities
In the mid 20th century, whole cities’ sewage systems safely and successfully used fish to treat and purify their water. Waste-fed fish ponds are a low-tech, cheap, and sustainable alternative to deal with our own shit — and to obtain high protein food in the process. ⌘ Read more

⤋ Read More

weewiki uses a custom org markup parser written in ANSI C to render the HTML. No emacs needed! my hope is to introduce a user-defined callback that can process these to allow for custom meta-commands.

⤋ Read More

I love it. I have a program that needs to processing about half a million records, which will take 3 days. The database that all those records are suppose to go to is acting up after I’ve just done 140K records.

⤋ Read More

Hot take: the existence of garden path sentences tells us that: 1) language isn’t so much a representation of meaning as a way of representing constraints for possible intepretation, and 2) we start processing before all the constraints have arrived

⤋ Read More