bash xfs_repair -L /dev/mapper/centos-root
bash service NetworkManager stop
### ADD RPM Forge repository | |
wget http://packages.sw.be/rpmforge-release/rpmforge-release-0.5.2-2.el6.rf.x86_64.rpm | |
rpm --import http://apt.sw.be/RPM-GPG-KEY.dag.txt | |
rpm -K rpmforge-release-0.5.2-2.el6.rf.*.rpm # Verifies the package | |
rpm -i rpmforge-release-0.5.2-2.el6.rf.*.rpm | |
### APACHE | |
# Dependencies |
If you’re trying to delete a very large number of files at one time (I deleted a directory with 485,000+ today), you will probably run into this error:
/bin/rm: Argument list too long.
The problem is that when you type something like “rm -rf ”, the “” is replaced with a list of every matching file, like “rm -rf file1 file2 file3 file4” and so on. There is a reletively small buffer of memory allocated to storing this list of arguments and if it is filled up, the shell will not execute the program. To get around this problem, a lot of people will use the find command to find every file and pass them one-by-one to the “rm” command like this:
find . -type f -exec rm -v {} \;
My problem is that I needed to delete 500,000 files and it was taking way too long.
<?php | |
define('API_KEY', 'YOUR API KEY HERE'); | |
define('SECRET', 'YOUR SECRET HERE'); | |
define('STORE_DOMAIN', 'YOUR STORE DOMAIN HERE'); | |
define('WEBHOOK_TARGET', 'your webhook URL here'); | |
ini_set('display_errors', 'On'); | |
ini_set('html_errors', 'Off'); | |
error_reporting(-1); |
#!/bin/bash | |
# Set these variables | |
USER_NAME="you@example.com" | |
API_KEY="XXXXXXXX_YOUR_DHQ_API_KEY_XXXXXXXXXXXXX" | |
START_REV='SOME_ARBITRARY_GIT_REVISION_HASH_TO_START_FROM' | |
DHQ_API_PROJ="your-project-shortname" | |
DHQ_BASE_URL="https://yoursite.deployhq.com/" | |
DHQ_SERVER_GROUP="YOUR_SERVER_GROUP_UUID" | |
DHQ_SERVER_USERNAME="your-server-username" |
#!/usr/bin/python | |
import os | |
import subprocess | |
import time | |
import yaml | |
import re | |
user_name = os.environ.get("DOCKERHUB_USER") |
NOTE: The service level and consistency feature is one of the most changed areas of the mymysqlnd_ms fork. Functionalities like server side read consistency and server side write consistency allow transparent migration to MySQL asyncronous clusters in almost all use cases with no or at most extremely small effort and application changes.
Different types of MySQL cluster solutions offer different service and data consistency levels to their users. Any asynchronous MySQL replication cluster offers eventual consistency by default. A read executed on an asynchronous slave may return current, stale or no data at all, depending
;################################################### | |
; NOTE | |
;#################################################### | |
; change your curl.ini on /etc/php.d/curl.ini | |
; | |
;This config was only tested on amazon AMI | |
;Please check if /etc/ssl/certs/ca-bundle.crt exists | |
; | |
;If you don't have the ca-bundle root certificate you | |
; can get this in |
$destination_path = 'C:\Users\dilli\Downloads\media_dump' | |
$connection_string = '[AZURE_STORAGE_CONNECTION_STRING]' | |
$storage_account = New-AzureStorageContext -ConnectionString $connection_string | |
$containers = Get-AzureStorageContainer -Context $storage_account | |
Write-Host 'Starting Storage Dump...' | |
foreach ($container in $containers) |