A voice signal $m(t)$ is in the frequency range $5\:kHz$ to $15\:kHz$. The signal is amplitude-modulated to generated an AM signal $f(t)=A\left(1+m(t)\right)\cos 2\pi f_{c}t,$ where $f_{c}=600\: kHz.$ The AM signal $f(t)$ is to be digitized and archived. This is done by first sampling $f(t)$ at $1.2$ times the Nyquist frequency, and then quantizing each sample using a $256$ – level quantizer. Finally, each quantized sample is binary coded using $K$ bits, where $K$ is the minimum number of bits required for the encoding. The rate, in Megabits per second (rounded off to $2$ decimal places), of the resulting stream of coded bits is ________ Mbps.