Motivation
It seems like a weird requirement to direct traffic out of intra-host network stack, traverse a switch, and bounce back to another interface: why one would ever need a intra-host communication with both worse latency and throughput? And if they does want a real network latency and thoughput, why not just put two ends of a protocol on two machines?
Well it happens, we were emulating for something. Two instances should be put under different filesystems and network namespaces: they must regard each other as separate nodes, while they must share access to the same hardware attached to the host machine, so we couldn’t put them on two baremetal machines.
Latency and thoughput noise injection was not good, either. Because TBD
Operation
We have a dual-port NIC with interfaces ens0
and ens1
recognized by host in default namespace, both are down, without address ever assigned.
- Separate them to isolated ns to avoid kernel shortcut
1 | sudo ip netns add ns0 |
- Verify they can ping each other via fabric
1 | sudo ip netns exec ns1 tcpdump -i ens1 icmp > logs.out & |
- Create
macvlan
s bridge with interfaces
1 | sudo ip netns exec ns0 ip link macvlan0 link ens0 type macvlan mode bridge |
- Create docker without network
1 | docker run -dit --name c0 --network none alpine sh |
- Move prepared
macvlan
s to corresponding container ns
1 | sudo ip link set macvlan0 netns $pid0 |
- Verify containers can ping each other via fabric:
1 | sudo ip netns exec ns1 tcpdump -i ens1 icmp > logs.out & |
But how many copies?
Should be only two, the same as a fabric communication: kernel -> fabric -> kernel.
Why it works?
TBD.
Lessons
TBD.
docker create macvlan network
create veth pairs