We’re using a MikroTik RB3011UiAS router with a gigabit (1G) link connected to our ISP’s NTU through ether1
The router does not perfom any Firewalling or NAT
We have configured traffic shaping on the MikroTik using “simple queues.” However, the ISP reports they’re still seeing policer drops on their end. Below is the shaper we are using on the Mikrotik.
How is the cabling done ? Do you use the SFP slot ?
Remember the RB3011 has an architecture where both groups of ports are interconnected to CPU-cores using a 1Gbits/sec channels. (1 for each CPU-core)
When you do queuing I’m pretty sure the CPU is involved so if you use 2 adjacent ports your probably have this result.
Packets travel from ingress port (your PC/server) → via CPU → back to egress port (NTU) so cutting speed in half worst case depending on how CPU is used for the queuing processes.
So a RB5009 would be a better choice for your setup I think, this has a full-duplex 10Gbits/s channel between the Marvell-switch chip and CPU-complex.
You can’t fast track simple queues. (not fast track is much slower)
You need to set a lower bucket size on the queues.
You can show system resources and watch the cpu when you run traffic through the link.
Thoughts.
Mostly want to limit outbound traffic correctly.
Inbound less of an issue (It is the ISP throttling the traffic)
My thought would be to test something like the following:
Rather than using simple queues Create a queue in queue tree and have ether1 (Wan port) as its parent interface.
Interface queues work with fast tracked traffic.
Have its packet mark being no mark.
Set the bucket size of the queue to 0.01 (or better 0.005).
(Approximately the max bucket size is a time <= time for 1035k byte at 980Mbps, smaller is more CPU)
Enable fast track (if you have previously disabled it)
Unless you are running fastpath (No firewall rules at all).
See how this goes.
If still running out of cpu, try downgrade to latest V6, same config.
We not using an SFP port. just using ether1 for WAN and ether2 for LAN so i could be an issue with the grouping?
Sure it could be that, try moving “ether2” for example to port 6/7/8/9/10 and see what happens.
I’m pretty sure throughput will for sure increase as you have a 1Gbps “link” to each CPU-core.
For 512 bytes no firewall (fastpath), will get around 1650M, but with a few rules,
will get 452M. Hopefully your scenario will be closer to no firewall.
But even 1650M is still well slower than 1Gb in both directions at the same time, so perhaps a device upgrade.
Apparently the switch chip will do some sort of ingress rate limiting in hardware, so you could maybe try that, but you will likely struggle to get
assistance with it.
Thanks you so much @rplant that new queue tree worked!!!
I am forever grateful
I do have one more issue though. we have another site with a similar issue. File transfers TO the DC (uploads) result in the following
40MBps to 60MBs without any shaper
~8MBs with the shaper
File transfers from the DC (downloads) are ~2MBps regardless of shaper/no shaper (but i suspect this is to do with a separate issue somewhere upstream)
Its the exact same model (3011), also with a 1Gig link to the ISP. The difference here is that there is considerably more latency to the DC (~25ms)