diff -r folder1 folder2 | grep -P '(^diff|^Only)'
diff -r folder1 folder2 | grep -P '(^diff|^Only)' | sed -r 's/^diff -r /vimdiff /'
-
https://links.infomee.fr/?Gc4iwQI may have to use this one, but issue number is freaking me out
-
https://github.com/s3fs-fuse/s3fs-fuseC'est vraiment se compliquer la vie pour pas grande chose
Et puis je m'en souviens jamais : en hiver on est GMT+1, en été on est GMT+2
-
https://links.infomee.fr/?X6cRbAsubnets = ec2.subnets.all()
subnets_sorted = sorted(subnets, key=lambda k: k.tags[next(index for (index, d) in enumerate(k.tags) if d["Key"] == "Name")]['Value'])
Well, my python level is not good enough to clearly understand this but my Google level was largely enough to build this
-
https://links.infomee.fr/?Fn--jwEasy one?
Not even close
-
https://aws.amazon.com/blogs/security/iam-policies-and-bucket-policies-and-acls-oh-my-controlling-access-to-s3-resources/openssl s_client -connect www.cyberciti.biz:443
-
https://www.cyberciti.biz/tips/debugging-ssl-communications-from-unix-shell-prompt.htmlAs the included diagram portrays, in a common workflow, the worker instance will consume messages sent to a specified Amazon SQS from another service (e.g.: a web server or another worker). These messages will be received by the worker via POST requests. This eliminates the necessity of configuring a worker as an always-on service, as well as having to add code for reading and consuming messages from an AWS SQS queue. In other words, the worker is implemented as a standard RESTful API/Service that will react to a message sent to it at an specific endpoint via a POST request. This is an awesome approach by Amazon to microservices and reactive design.
The conversion of the SQS message to a POST request is executed by what AWS calls the "SQS Daemon" or "Sqsd". This is a simple daemon they pre-install in the worker tier instances that is constantly monitoring an specific AWS SQS queue (provided by configuration) for new messages. When new messages arrive, it constructs a POST request and sends it to a specific endpoint (also provided via configuration). If the endpoint consumes it without errors and returns a 2** HTTP Code in its response, the "Sqsd" deletes the message from the queue to signal that its consumption was successful.
However, even though this approach is extremely powerful, Amazon does not provide the code of this daemon as open source. Therefore, we have reproduced its behavior by creating our own version of the "Sqsd" free for everyone to use. Moreover, we have provided lots of customization and configuration properties so that it can be molded to your specific use cases.
-
https://github.com/mozart-analytics/sqsdDon't give s3 full access policy to your app user
Prefer to allow access only for specific bucket
I wonder what's the best choice : create managed policy or simply use inline policy. I got a 1 to 1 relationship between my app-users and bucket so... inline policy looks good here
-
https://aws.amazon.com/blogs/security/writing-iam-policies-how-to-grant-access-to-an-amazon-s3-bucket/