Each of these commands will run an ad hoc http static server in your current (or specified) directory, available at http://localhost:8000. Use this power wisely.
$ python -m SimpleHTTPServer 8000
Each of these commands will run an ad hoc http static server in your current (or specified) directory, available at http://localhost:8000. Use this power wisely.
$ python -m SimpleHTTPServer 8000
To limit a CPU to a certain C-state, you can pass the processor.max_cstate=X
option in the kernel
line of /boot/grub/grub.conf
.
Here we limit the system to only C-State 1:
kernel /vmlinuz-2.6.18-371.1.2.el5 ... processor.max_cstate=1
On some systems, the kernel can override the BIOS setting, and the parameter intel_idle.max_cstate=0
may be required to ensure sleep states are not entered:
$/
artifacts/
build/
docs/
lib/
packages/
samples/
src/
tests/
Please see other gists for updated information. https://gist.github.com/XenoPanther/15d8fad49fbd51c6bd946f2974084ef8
[ Update 2020-05-31: I won't be maintaining this page or responding to comments anymore (except for perhaps a few exceptional occasions). ]
Most of the terminal emulators auto-detect when a URL appears onscreen and allow to conveniently open them (e.g. via Ctrl+click or Cmd+click, or the right click menu).
It was, however, not possible until now for arbitrary text to point to URLs, just as on webpages.
#!/usr/bin/env sh | |
## I don't have any devices running Android OS now. These settings may have been outdated but I cannot verify it. | |
adb shell settings put global captive_portal_fallback_url http://g.cn/generate_204 | |
adb shell settings put global captive_portal_http_url http://www.google.cn/generate_204 | |
adb shell settings put global captive_portal_https_url https://www.google.cn/generate_204 | |
adb shell settings put global captive_portal_other_fallback_urls http://www.qualcomm.cn/generate_204 |
Here I will describe a simple configuration of the slurm management tool for launching jobs in a really simplistic cluster. I will assume the following configuration: a main node (for me it is an Arch Linux distribution) and 3 compute nodes (for me compute nodes are Debian VMs). I also assume there is ping access between the nodes and some sort of mechanism for you to know the IP of each node at all times (most basic should be a local NAT with static IPs)
Slurm management tool work on a set of nodes, one of which is considered the master node, and has the slurmctld
daemon running; all other compute nodes have the slurmd
daemon. All communications are authenticated via the munge
service and all nodes need to share the same authentication key. Slurm by default holds a journal of activities in a directory configured in the slurm.conf
file, however a Database management system can be set. All in all what we will try to do is:
The official guide for setting up Kubernetes using kubeadm
works well for clusters of one architecture. But, the main problem that crops up is the kube-proxy
image defaults to the architecture of the master node (where kubeadm
was run in the first place).
This causes issues when arm
nodes join the cluster, as they will try to execute the amd64
version of kube-proxy
, and will fail.
It turns out that the pod running kube-proxy
is configured using a DaemonSet. With a small edit to the configuration, it's possible to create multiple DaemonSets—one for each architecture.
Follow the instructions at https://kubernetes.io/docs/setup/independent/create-cluster-kubeadm/ for setting up the master node. I've been using Weave Net as the network plugin; it see
Based on https://gist.github.com/jjvillavicencio/4e3615a8219bb1a17c81c4541c6c317d
server side: ib_send_bw -d mlx5_2 -i 1 -x 0 | |
mlx5_2 is the PF or VF dev name, | |
-x is needed for SR-IOV | |
client side: ib_send_bw -d mlx5_4 -i 1 10.56.217.100 -x 0 | |
mlx5_4 is the VF dev name inside pod, | |
10.56.217.100 is the server side IP address of VF or PF. | |
-x is needed for SR-IOV. |